var/home/core/zuul-output/0000755000175000017500000000000015066536301014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015066542723015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003655667015066542716017733 0ustar rootrootSep 29 17:09:30 crc systemd[1]: Starting Kubernetes Kubelet... Sep 29 17:09:30 crc restorecon[4538]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:30 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Sep 29 17:09:31 crc restorecon[4538]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Sep 29 17:09:31 crc kubenswrapper[4667]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 17:09:31 crc kubenswrapper[4667]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Sep 29 17:09:31 crc kubenswrapper[4667]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 17:09:31 crc kubenswrapper[4667]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 17:09:31 crc kubenswrapper[4667]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 29 17:09:31 crc kubenswrapper[4667]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.689364 4667 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693398 4667 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693421 4667 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693426 4667 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693430 4667 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693434 4667 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693439 4667 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693445 4667 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693449 4667 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693452 4667 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693456 4667 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693459 4667 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693463 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693466 4667 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693469 4667 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693472 4667 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693475 4667 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693479 4667 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693484 4667 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693487 4667 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693491 4667 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693494 4667 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693498 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693501 4667 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693505 4667 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693509 4667 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693512 4667 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693515 4667 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693518 4667 feature_gate.go:330] unrecognized feature gate: Example Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693522 4667 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693525 4667 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693529 4667 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693532 4667 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693535 4667 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693545 4667 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693549 4667 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693553 4667 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693556 4667 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693561 4667 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693564 4667 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693568 4667 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693572 4667 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693576 4667 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693579 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693583 4667 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693587 4667 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693591 4667 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693595 4667 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693599 4667 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693602 4667 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693605 4667 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693609 4667 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693614 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693618 4667 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693621 4667 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693624 4667 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693627 4667 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693630 4667 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693633 4667 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693637 4667 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693640 4667 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693643 4667 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693646 4667 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693649 4667 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693653 4667 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693657 4667 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693662 4667 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693665 4667 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693668 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693671 4667 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693674 4667 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.693677 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693798 4667 flags.go:64] FLAG: --address="0.0.0.0" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693809 4667 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693818 4667 flags.go:64] FLAG: --anonymous-auth="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693823 4667 flags.go:64] FLAG: --application-metrics-count-limit="100" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693830 4667 flags.go:64] FLAG: --authentication-token-webhook="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693833 4667 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693839 4667 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693858 4667 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693862 4667 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693866 4667 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693871 4667 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693875 4667 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693880 4667 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693884 4667 flags.go:64] FLAG: --cgroup-root="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693888 4667 flags.go:64] FLAG: --cgroups-per-qos="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693891 4667 flags.go:64] FLAG: --client-ca-file="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693895 4667 flags.go:64] FLAG: --cloud-config="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693899 4667 flags.go:64] FLAG: --cloud-provider="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693903 4667 flags.go:64] FLAG: --cluster-dns="[]" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693907 4667 flags.go:64] FLAG: --cluster-domain="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693911 4667 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693915 4667 flags.go:64] FLAG: --config-dir="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693918 4667 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693923 4667 flags.go:64] FLAG: --container-log-max-files="5" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693928 4667 flags.go:64] FLAG: --container-log-max-size="10Mi" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693932 4667 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693936 4667 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693941 4667 flags.go:64] FLAG: --containerd-namespace="k8s.io" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693945 4667 flags.go:64] FLAG: --contention-profiling="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693949 4667 flags.go:64] FLAG: --cpu-cfs-quota="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693952 4667 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693956 4667 flags.go:64] FLAG: --cpu-manager-policy="none" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693960 4667 flags.go:64] FLAG: --cpu-manager-policy-options="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693965 4667 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693968 4667 flags.go:64] FLAG: --enable-controller-attach-detach="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693972 4667 flags.go:64] FLAG: --enable-debugging-handlers="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693976 4667 flags.go:64] FLAG: --enable-load-reader="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693980 4667 flags.go:64] FLAG: --enable-server="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693983 4667 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693989 4667 flags.go:64] FLAG: --event-burst="100" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693993 4667 flags.go:64] FLAG: --event-qps="50" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.693997 4667 flags.go:64] FLAG: --event-storage-age-limit="default=0" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694001 4667 flags.go:64] FLAG: --event-storage-event-limit="default=0" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694004 4667 flags.go:64] FLAG: --eviction-hard="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694009 4667 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694013 4667 flags.go:64] FLAG: --eviction-minimum-reclaim="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694016 4667 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694021 4667 flags.go:64] FLAG: --eviction-soft="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694025 4667 flags.go:64] FLAG: --eviction-soft-grace-period="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694029 4667 flags.go:64] FLAG: --exit-on-lock-contention="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694032 4667 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694036 4667 flags.go:64] FLAG: --experimental-mounter-path="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694040 4667 flags.go:64] FLAG: --fail-cgroupv1="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694044 4667 flags.go:64] FLAG: --fail-swap-on="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694048 4667 flags.go:64] FLAG: --feature-gates="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694058 4667 flags.go:64] FLAG: --file-check-frequency="20s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694062 4667 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694066 4667 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694071 4667 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694075 4667 flags.go:64] FLAG: --healthz-port="10248" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694079 4667 flags.go:64] FLAG: --help="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694083 4667 flags.go:64] FLAG: --hostname-override="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694086 4667 flags.go:64] FLAG: --housekeeping-interval="10s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694090 4667 flags.go:64] FLAG: --http-check-frequency="20s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694094 4667 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694097 4667 flags.go:64] FLAG: --image-credential-provider-config="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694101 4667 flags.go:64] FLAG: --image-gc-high-threshold="85" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694105 4667 flags.go:64] FLAG: --image-gc-low-threshold="80" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694108 4667 flags.go:64] FLAG: --image-service-endpoint="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694112 4667 flags.go:64] FLAG: --kernel-memcg-notification="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694116 4667 flags.go:64] FLAG: --kube-api-burst="100" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694119 4667 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694124 4667 flags.go:64] FLAG: --kube-api-qps="50" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694127 4667 flags.go:64] FLAG: --kube-reserved="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694132 4667 flags.go:64] FLAG: --kube-reserved-cgroup="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694135 4667 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694139 4667 flags.go:64] FLAG: --kubelet-cgroups="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694143 4667 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694147 4667 flags.go:64] FLAG: --lock-file="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694151 4667 flags.go:64] FLAG: --log-cadvisor-usage="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694155 4667 flags.go:64] FLAG: --log-flush-frequency="5s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694159 4667 flags.go:64] FLAG: --log-json-info-buffer-size="0" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694165 4667 flags.go:64] FLAG: --log-json-split-stream="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694170 4667 flags.go:64] FLAG: --log-text-info-buffer-size="0" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694174 4667 flags.go:64] FLAG: --log-text-split-stream="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694177 4667 flags.go:64] FLAG: --logging-format="text" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694181 4667 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694185 4667 flags.go:64] FLAG: --make-iptables-util-chains="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694189 4667 flags.go:64] FLAG: --manifest-url="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694193 4667 flags.go:64] FLAG: --manifest-url-header="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694199 4667 flags.go:64] FLAG: --max-housekeeping-interval="15s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694203 4667 flags.go:64] FLAG: --max-open-files="1000000" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694208 4667 flags.go:64] FLAG: --max-pods="110" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694212 4667 flags.go:64] FLAG: --maximum-dead-containers="-1" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694215 4667 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694219 4667 flags.go:64] FLAG: --memory-manager-policy="None" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694223 4667 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694227 4667 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694231 4667 flags.go:64] FLAG: --node-ip="192.168.126.11" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694235 4667 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694245 4667 flags.go:64] FLAG: --node-status-max-images="50" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694249 4667 flags.go:64] FLAG: --node-status-update-frequency="10s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694253 4667 flags.go:64] FLAG: --oom-score-adj="-999" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694256 4667 flags.go:64] FLAG: --pod-cidr="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694260 4667 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694266 4667 flags.go:64] FLAG: --pod-manifest-path="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694270 4667 flags.go:64] FLAG: --pod-max-pids="-1" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694274 4667 flags.go:64] FLAG: --pods-per-core="0" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694278 4667 flags.go:64] FLAG: --port="10250" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694281 4667 flags.go:64] FLAG: --protect-kernel-defaults="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694285 4667 flags.go:64] FLAG: --provider-id="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694289 4667 flags.go:64] FLAG: --qos-reserved="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694293 4667 flags.go:64] FLAG: --read-only-port="10255" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694296 4667 flags.go:64] FLAG: --register-node="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694300 4667 flags.go:64] FLAG: --register-schedulable="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694303 4667 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694310 4667 flags.go:64] FLAG: --registry-burst="10" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694314 4667 flags.go:64] FLAG: --registry-qps="5" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694318 4667 flags.go:64] FLAG: --reserved-cpus="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694322 4667 flags.go:64] FLAG: --reserved-memory="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694327 4667 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694331 4667 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694335 4667 flags.go:64] FLAG: --rotate-certificates="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694356 4667 flags.go:64] FLAG: --rotate-server-certificates="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694360 4667 flags.go:64] FLAG: --runonce="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694364 4667 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694368 4667 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694372 4667 flags.go:64] FLAG: --seccomp-default="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694379 4667 flags.go:64] FLAG: --serialize-image-pulls="true" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694383 4667 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694387 4667 flags.go:64] FLAG: --storage-driver-db="cadvisor" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694391 4667 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694395 4667 flags.go:64] FLAG: --storage-driver-password="root" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694399 4667 flags.go:64] FLAG: --storage-driver-secure="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694403 4667 flags.go:64] FLAG: --storage-driver-table="stats" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694408 4667 flags.go:64] FLAG: --storage-driver-user="root" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694412 4667 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694416 4667 flags.go:64] FLAG: --sync-frequency="1m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694420 4667 flags.go:64] FLAG: --system-cgroups="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694423 4667 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694429 4667 flags.go:64] FLAG: --system-reserved-cgroup="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694433 4667 flags.go:64] FLAG: --tls-cert-file="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694437 4667 flags.go:64] FLAG: --tls-cipher-suites="[]" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694441 4667 flags.go:64] FLAG: --tls-min-version="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694445 4667 flags.go:64] FLAG: --tls-private-key-file="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694448 4667 flags.go:64] FLAG: --topology-manager-policy="none" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694452 4667 flags.go:64] FLAG: --topology-manager-policy-options="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694456 4667 flags.go:64] FLAG: --topology-manager-scope="container" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694460 4667 flags.go:64] FLAG: --v="2" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694465 4667 flags.go:64] FLAG: --version="false" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694470 4667 flags.go:64] FLAG: --vmodule="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694475 4667 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.694479 4667 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694580 4667 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694585 4667 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694591 4667 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694594 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694598 4667 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694602 4667 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694605 4667 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694610 4667 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694613 4667 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694617 4667 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694621 4667 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694624 4667 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694628 4667 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694633 4667 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694636 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694640 4667 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694643 4667 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694647 4667 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694651 4667 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694655 4667 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694659 4667 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694663 4667 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694666 4667 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694669 4667 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694673 4667 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694676 4667 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694680 4667 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694683 4667 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694686 4667 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694689 4667 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694693 4667 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694696 4667 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694699 4667 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694702 4667 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694707 4667 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694710 4667 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694714 4667 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694717 4667 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694720 4667 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694725 4667 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694728 4667 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694732 4667 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694735 4667 feature_gate.go:330] unrecognized feature gate: Example Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694738 4667 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694742 4667 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694745 4667 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694748 4667 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694752 4667 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694757 4667 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694761 4667 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694765 4667 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694769 4667 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694773 4667 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694776 4667 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694780 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694783 4667 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694786 4667 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694789 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694793 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694796 4667 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694799 4667 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694802 4667 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694807 4667 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694811 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694814 4667 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694817 4667 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694822 4667 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694825 4667 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694829 4667 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694832 4667 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.694835 4667 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.695211 4667 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.702916 4667 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.702950 4667 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703035 4667 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703050 4667 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703054 4667 feature_gate.go:330] unrecognized feature gate: Example Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703058 4667 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703062 4667 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703066 4667 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703070 4667 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703074 4667 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703077 4667 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703080 4667 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703084 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703087 4667 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703090 4667 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703093 4667 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703097 4667 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703102 4667 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703105 4667 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703109 4667 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703112 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703115 4667 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703118 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703121 4667 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703124 4667 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703128 4667 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703131 4667 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703134 4667 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703137 4667 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703140 4667 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703143 4667 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703147 4667 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703150 4667 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703153 4667 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703156 4667 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703160 4667 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703164 4667 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703167 4667 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703171 4667 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703175 4667 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703179 4667 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703183 4667 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703187 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703191 4667 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703195 4667 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703199 4667 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703202 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703206 4667 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703209 4667 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703212 4667 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703215 4667 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703219 4667 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703222 4667 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703225 4667 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703229 4667 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703232 4667 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703235 4667 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703240 4667 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703244 4667 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703248 4667 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703253 4667 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703257 4667 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703261 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703264 4667 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703268 4667 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703272 4667 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703277 4667 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703280 4667 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703283 4667 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703287 4667 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703290 4667 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703295 4667 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703299 4667 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.703305 4667 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703436 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703445 4667 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703448 4667 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703451 4667 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703454 4667 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703457 4667 feature_gate.go:330] unrecognized feature gate: NewOLM Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703461 4667 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703464 4667 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703467 4667 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703470 4667 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703473 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703476 4667 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703480 4667 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703483 4667 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703486 4667 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703489 4667 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703492 4667 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703496 4667 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703500 4667 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703506 4667 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703509 4667 feature_gate.go:330] unrecognized feature gate: Example Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703513 4667 feature_gate.go:330] unrecognized feature gate: InsightsConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703516 4667 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703520 4667 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703523 4667 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703527 4667 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703530 4667 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703534 4667 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703537 4667 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703540 4667 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703544 4667 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703547 4667 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703551 4667 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703555 4667 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703559 4667 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703563 4667 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703567 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703571 4667 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703575 4667 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703579 4667 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703582 4667 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703586 4667 feature_gate.go:330] unrecognized feature gate: GatewayAPI Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703590 4667 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703593 4667 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703596 4667 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703599 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703602 4667 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703606 4667 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703610 4667 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703613 4667 feature_gate.go:330] unrecognized feature gate: OVNObservability Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703616 4667 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703620 4667 feature_gate.go:330] unrecognized feature gate: SignatureStores Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703623 4667 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703626 4667 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703629 4667 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703632 4667 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703636 4667 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703639 4667 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703642 4667 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703645 4667 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703648 4667 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703651 4667 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703654 4667 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703657 4667 feature_gate.go:330] unrecognized feature gate: PlatformOperators Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703661 4667 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703664 4667 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703667 4667 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703671 4667 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703675 4667 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703679 4667 feature_gate.go:330] unrecognized feature gate: PinnedImages Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.703683 4667 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.703689 4667 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.704301 4667 server.go:940] "Client rotation is on, will bootstrap in background" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.707012 4667 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.707097 4667 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.707881 4667 server.go:997] "Starting client certificate rotation" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.707905 4667 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.708086 4667 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-27 01:56:49.963569743 +0000 UTC Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.708161 4667 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1400h47m18.25541084s for next certificate rotation Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.719993 4667 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.722610 4667 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.731944 4667 log.go:25] "Validated CRI v1 runtime API" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.750796 4667 log.go:25] "Validated CRI v1 image API" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.752241 4667 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.755472 4667 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-09-29-17-06-19-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.755493 4667 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm:{mountpoint:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm major:0 minor:42 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:50 fsType:tmpfs blockSize:0} overlay_0-43:{mountpoint:/var/lib/containers/storage/overlay/94b752e0a51c0134b00ddef6dc7a933a9d7c1d9bdc88a18dae4192a0d557d623/merged major:0 minor:43 fsType:overlay blockSize:0}] Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.768730 4667 manager.go:217] Machine: {Timestamp:2025-09-29 17:09:31.767021726 +0000 UTC m=+0.264868515 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2445406 MemoryCapacity:33654116352 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:d5a94666-8121-4bfb-8540-72964a1282ac BootID:f7158ada-47b4-429a-bd74-dd92a5b97fd6 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827056128 Type:vfs Inodes:4108168 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm DeviceMajor:0 DeviceMinor:42 Capacity:65536000 Type:vfs Inodes:4108168 HasInodes:true} {Device:overlay_0-43 DeviceMajor:0 DeviceMinor:43 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:50 Capacity:1073741824 Type:vfs Inodes:4108168 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e1:f2:5f Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:enp3s0 MacAddress:fa:16:3e:e1:f2:5f Speed:-1 Mtu:1500} {Name:enp7s0 MacAddress:fa:16:3e:23:35:5d Speed:-1 Mtu:1440} {Name:enp7s0.20 MacAddress:52:54:00:13:0e:e5 Speed:-1 Mtu:1436} {Name:enp7s0.21 MacAddress:52:54:00:1c:7f:8a Speed:-1 Mtu:1436} {Name:enp7s0.22 MacAddress:52:54:00:c8:3c:11 Speed:-1 Mtu:1436} {Name:eth10 MacAddress:fe:28:03:97:f5:f2 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:1e:21:b4:4d:f1:d4 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654116352 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:65536 Type:Data Level:1} {Id:0 Size:65536 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:65536 Type:Data Level:1} {Id:1 Size:65536 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:65536 Type:Data Level:1} {Id:10 Size:65536 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:65536 Type:Data Level:1} {Id:11 Size:65536 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:65536 Type:Data Level:1} {Id:2 Size:65536 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:65536 Type:Data Level:1} {Id:3 Size:65536 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:65536 Type:Data Level:1} {Id:4 Size:65536 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:65536 Type:Data Level:1} {Id:5 Size:65536 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:65536 Type:Data Level:1} {Id:6 Size:65536 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:65536 Type:Data Level:1} {Id:7 Size:65536 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:65536 Type:Data Level:1} {Id:8 Size:65536 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:65536 Type:Data Level:1} {Id:9 Size:65536 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.768917 4667 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.769039 4667 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.769709 4667 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.769894 4667 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.769928 4667 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.770141 4667 topology_manager.go:138] "Creating topology manager with none policy" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.770150 4667 container_manager_linux.go:303] "Creating device plugin manager" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.770464 4667 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.770491 4667 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.770782 4667 state_mem.go:36] "Initialized new in-memory state store" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.770929 4667 server.go:1245] "Using root directory" path="/var/lib/kubelet" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.774550 4667 kubelet.go:418] "Attempting to sync node with API server" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.774569 4667 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.774591 4667 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.774603 4667 kubelet.go:324] "Adding apiserver pod source" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.774614 4667 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.777053 4667 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.777691 4667 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.779018 4667 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.779313 4667 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.779324 4667 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.779404 4667 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.779421 4667 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780000 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780023 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780033 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780040 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780050 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780056 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780062 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780073 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780080 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780086 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780111 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780118 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.780668 4667 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.781193 4667 server.go:1280] "Started kubelet" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.782609 4667 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.783419 4667 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 29 17:09:31 crc systemd[1]: Started Kubernetes Kubelet. Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.784157 4667 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.784497 4667 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.785124 4667 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.785180 4667 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.785603 4667 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.785621 4667 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 11:40:04.329856408 +0000 UTC Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.785660 4667 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1890h30m32.544198146s for next certificate rotation Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.786136 4667 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="200ms" Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.786234 4667 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.785607 4667 volume_manager.go:287] "The desired_state_of_world populator starts" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.786500 4667 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.786473 4667 server.go:460] "Adding debug handlers to kubelet server" Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.786626 4667 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.786676 4667 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.786778 4667 factory.go:55] Registering systemd factory Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.786861 4667 factory.go:221] Registration of the systemd container factory successfully Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.788516 4667 factory.go:153] Registering CRI-O factory Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.788543 4667 factory.go:221] Registration of the crio container factory successfully Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.788599 4667 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.788622 4667 factory.go:103] Registering Raw factory Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.788640 4667 manager.go:1196] Started watching for new ooms in manager Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.789188 4667 manager.go:319] Starting recovery of all containers Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.787663 4667 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 192.168.26.120:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1869cff2f0084c09 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-09-29 17:09:31.781000201 +0000 UTC m=+0.278846961,LastTimestamp:2025-09-29 17:09:31.781000201 +0000 UTC m=+0.278846961,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794697 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794754 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794765 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794776 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794784 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794792 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794800 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794808 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794822 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794830 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794852 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794873 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794882 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794892 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794901 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794908 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794915 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794922 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794931 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794940 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794951 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794958 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794966 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794975 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794982 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794989 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.794998 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795007 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795015 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795024 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795044 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795053 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795063 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795071 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795079 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795087 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795117 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795128 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795137 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795145 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795152 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795162 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795172 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795180 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795189 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795197 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795205 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795214 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795221 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795230 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795244 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795252 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795263 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795272 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795281 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795289 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795297 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795306 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795313 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795321 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795328 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795337 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795357 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795366 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795375 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795384 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795391 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795400 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795408 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795415 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795423 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795431 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795438 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795446 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795453 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795461 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795483 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795492 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795500 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795508 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795516 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795523 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795543 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795552 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795560 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795567 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795574 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795582 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795590 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795598 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795606 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795614 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795622 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795630 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795637 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795647 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795654 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795661 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795669 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795676 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795683 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795690 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795697 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795705 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795716 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795726 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795734 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795743 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795754 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795764 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795772 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795780 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795788 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795797 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795805 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795814 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795822 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795829 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795837 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795886 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795895 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795905 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795914 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795922 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795932 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795941 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795949 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795957 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795966 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795980 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795989 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.795997 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796004 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796012 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796021 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796029 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796036 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796045 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796053 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796061 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796069 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796078 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796086 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796093 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796101 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796108 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796116 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796125 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796132 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796139 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796147 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796155 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796163 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796170 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796177 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796185 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796193 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796202 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796210 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796219 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796227 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796235 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796242 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796252 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796259 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796266 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796274 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796282 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796289 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796296 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796305 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796312 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796322 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796330 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796338 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796359 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796367 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796375 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796382 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796390 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796398 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796405 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796412 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796439 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796448 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796455 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796462 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796470 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.796478 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797280 4667 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797299 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797309 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797318 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797327 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797335 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797343 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797365 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797373 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797380 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797388 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797395 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797403 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797410 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797417 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797431 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797441 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797448 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797455 4667 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797462 4667 reconstruct.go:97] "Volume reconstruction finished" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.797469 4667 reconciler.go:26] "Reconciler: start to sync state" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.808117 4667 manager.go:324] Recovery completed Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.812982 4667 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.814392 4667 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.814434 4667 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.814469 4667 kubelet.go:2335] "Starting kubelet main sync loop" Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.814510 4667 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 29 17:09:31 crc kubenswrapper[4667]: W0929 17:09:31.815649 4667 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.815717 4667 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.815782 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.816709 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.816742 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.816753 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.817380 4667 cpu_manager.go:225] "Starting CPU manager" policy="none" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.817406 4667 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.817424 4667 state_mem.go:36] "Initialized new in-memory state store" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.822013 4667 policy_none.go:49] "None policy: Start" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.822530 4667 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.822557 4667 state_mem.go:35] "Initializing new in-memory state store" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.865615 4667 manager.go:334] "Starting Device Plugin manager" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.865653 4667 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.865666 4667 server.go:79] "Starting device plugin registration server" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.865935 4667 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.865954 4667 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.866100 4667 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.866184 4667 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.866196 4667 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.873044 4667 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.914902 4667 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.915013 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.915881 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.915921 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.915931 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.916090 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.916307 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.916360 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.916836 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.916878 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.916890 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.916953 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917032 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917082 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917426 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917457 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917466 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917539 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917554 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917563 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917635 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917721 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917742 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.917988 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918022 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918031 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918234 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918277 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918301 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918235 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918367 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918378 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918525 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918657 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.918693 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919030 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919052 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919061 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919184 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919203 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919572 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919590 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919599 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919629 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919642 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.919650 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.966668 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.968939 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.968983 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.968993 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.969020 4667 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.969601 4667 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Sep 29 17:09:31 crc kubenswrapper[4667]: E0929 17:09:31.987184 4667 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="400ms" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999566 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999610 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999626 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999645 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999660 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999675 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999688 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999702 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999762 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999807 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999824 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999863 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999881 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999903 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:31 crc kubenswrapper[4667]: I0929 17:09:31.999916 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101414 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101488 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101508 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101523 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101537 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101550 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101565 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101581 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101596 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101600 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101609 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101624 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101644 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101659 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101669 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101676 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101689 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101691 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101709 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101722 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101730 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101737 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101750 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101769 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101787 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101629 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101811 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101827 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101877 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.101898 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.170446 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.171696 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.171795 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.171896 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.172018 4667 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 17:09:32 crc kubenswrapper[4667]: E0929 17:09:32.172435 4667 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.236320 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.253362 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: W0929 17:09:32.256283 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2aba2ace3922bfefc8a573fa612f86a4b4e0cf203752eebf6d9e53d41df820f6 WatchSource:0}: Error finding container 2aba2ace3922bfefc8a573fa612f86a4b4e0cf203752eebf6d9e53d41df820f6: Status 404 returned error can't find the container with id 2aba2ace3922bfefc8a573fa612f86a4b4e0cf203752eebf6d9e53d41df820f6 Sep 29 17:09:32 crc kubenswrapper[4667]: W0929 17:09:32.266206 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-191674a94e2631fb238999d6f5075cc12e7fd24f8686f6865f469f0f0e4deaa6 WatchSource:0}: Error finding container 191674a94e2631fb238999d6f5075cc12e7fd24f8686f6865f469f0f0e4deaa6: Status 404 returned error can't find the container with id 191674a94e2631fb238999d6f5075cc12e7fd24f8686f6865f469f0f0e4deaa6 Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.271738 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.276461 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.279817 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:32 crc kubenswrapper[4667]: W0929 17:09:32.281893 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-32672c2b211e4b3bf237e5ea3a4313457534e8dd51a23db37ec39f9bd5b250d5 WatchSource:0}: Error finding container 32672c2b211e4b3bf237e5ea3a4313457534e8dd51a23db37ec39f9bd5b250d5: Status 404 returned error can't find the container with id 32672c2b211e4b3bf237e5ea3a4313457534e8dd51a23db37ec39f9bd5b250d5 Sep 29 17:09:32 crc kubenswrapper[4667]: W0929 17:09:32.288431 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-ae377966fda862a09569fa58af04d01effd0d82ba1bacbd40f31644ee50ccd0a WatchSource:0}: Error finding container ae377966fda862a09569fa58af04d01effd0d82ba1bacbd40f31644ee50ccd0a: Status 404 returned error can't find the container with id ae377966fda862a09569fa58af04d01effd0d82ba1bacbd40f31644ee50ccd0a Sep 29 17:09:32 crc kubenswrapper[4667]: W0929 17:09:32.289456 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-717f1c09517c5f506518bff560745cff6bacee1ace59ba23be31b0a841632e88 WatchSource:0}: Error finding container 717f1c09517c5f506518bff560745cff6bacee1ace59ba23be31b0a841632e88: Status 404 returned error can't find the container with id 717f1c09517c5f506518bff560745cff6bacee1ace59ba23be31b0a841632e88 Sep 29 17:09:32 crc kubenswrapper[4667]: E0929 17:09:32.387742 4667 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="800ms" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.573127 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.574182 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.574205 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.574214 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.574232 4667 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 17:09:32 crc kubenswrapper[4667]: E0929 17:09:32.574526 4667 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Sep 29 17:09:32 crc kubenswrapper[4667]: W0929 17:09:32.739798 4667 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:32 crc kubenswrapper[4667]: E0929 17:09:32.739923 4667 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.784132 4667 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:32 crc kubenswrapper[4667]: W0929 17:09:32.808029 4667 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:32 crc kubenswrapper[4667]: E0929 17:09:32.808094 4667 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.819241 4667 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767" exitCode=0 Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.819321 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767"} Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.819442 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"717f1c09517c5f506518bff560745cff6bacee1ace59ba23be31b0a841632e88"} Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.820186 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.822698 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.822732 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.822743 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.823617 4667 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1fe7aa4a24221e184afdd808a3a04146b5fb2ef77c3720ca69a1358aff749618" exitCode=0 Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.823703 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1fe7aa4a24221e184afdd808a3a04146b5fb2ef77c3720ca69a1358aff749618"} Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.823760 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ae377966fda862a09569fa58af04d01effd0d82ba1bacbd40f31644ee50ccd0a"} Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.823886 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.824220 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.824575 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.824599 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.824609 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.825510 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.825535 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.825544 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.825633 4667 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b7e08e45811bd7a6442cc15a3212ba001bc4ef2df1ddda320c9a139f7fd01fdd" exitCode=0 Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.825696 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b7e08e45811bd7a6442cc15a3212ba001bc4ef2df1ddda320c9a139f7fd01fdd"} Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.825714 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"32672c2b211e4b3bf237e5ea3a4313457534e8dd51a23db37ec39f9bd5b250d5"} Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.825762 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.826490 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.826509 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.826518 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.827490 4667 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36" exitCode=0 Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.827534 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36"} Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.827618 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"191674a94e2631fb238999d6f5075cc12e7fd24f8686f6865f469f0f0e4deaa6"} Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.827689 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.828302 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.828328 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.828338 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.829304 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c"} Sep 29 17:09:32 crc kubenswrapper[4667]: I0929 17:09:32.829328 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2aba2ace3922bfefc8a573fa612f86a4b4e0cf203752eebf6d9e53d41df820f6"} Sep 29 17:09:33 crc kubenswrapper[4667]: W0929 17:09:33.016708 4667 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:33 crc kubenswrapper[4667]: E0929 17:09:33.016795 4667 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Sep 29 17:09:33 crc kubenswrapper[4667]: W0929 17:09:33.132266 4667 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.120:6443: connect: connection refused Sep 29 17:09:33 crc kubenswrapper[4667]: E0929 17:09:33.132361 4667 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.120:6443: connect: connection refused" logger="UnhandledError" Sep 29 17:09:33 crc kubenswrapper[4667]: E0929 17:09:33.188902 4667 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.120:6443: connect: connection refused" interval="1.6s" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.375333 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.376571 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.376626 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.376637 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.376665 4667 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 17:09:33 crc kubenswrapper[4667]: E0929 17:09:33.377172 4667 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.120:6443: connect: connection refused" node="crc" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.838778 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.838817 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.838829 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.838861 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.838870 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.838957 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.839625 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.839660 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.839670 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.840102 4667 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b98ed387e17159598d2c11525b5534226b01ac0f7beb08e43aed9d93024de185" exitCode=0 Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.840142 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b98ed387e17159598d2c11525b5534226b01ac0f7beb08e43aed9d93024de185"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.840295 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.840901 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.840924 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.840933 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.842484 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e1c15dc05de758c79f9d53e0d55d88187ddbf28778b9d70a52cd19d31a3d47a0"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.842543 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.843058 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.843091 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.843100 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.847675 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.847699 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.847709 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.847775 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.849144 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.849174 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.849191 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.850440 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.850473 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.850483 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189"} Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.850531 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.851032 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.851056 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:33 crc kubenswrapper[4667]: I0929 17:09:33.851064 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.323500 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.381158 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.716772 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.856760 4667 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6ae461746f91f42609f4b85cf93ffcfd68aa9b865190a44af69532e600a55f46" exitCode=0 Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.856884 4667 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.856919 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.856930 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.856880 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6ae461746f91f42609f4b85cf93ffcfd68aa9b865190a44af69532e600a55f46"} Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.856931 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.857069 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.859282 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.859323 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.859339 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.859394 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.859501 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.859532 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.859532 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.859565 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.859583 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.860813 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.860832 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.860858 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.978134 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.978928 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.978956 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.978965 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:34 crc kubenswrapper[4667]: I0929 17:09:34.978983 4667 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.862392 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"57073aea9b43bae01cf9e3f0e23390e2550bc3942f07cf94bdb9b92f12d4a9ca"} Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.862429 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2f9d89da318419cde5f78e062ae14168f716c169fb0bb7a2a7468974021ff75b"} Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.862440 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6bbfa2510696eb81866c2158c7eb7102a0c591f568ed85b48ba594d769ddd98d"} Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.862448 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bb7c287f124ce6b9696a6ae200279ba2e4509201c42176c5f1c21d78f948ea3a"} Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.862455 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"43d6f02c341c7866bad9a9a485185b47b00d93f8d0b2f289d27f1ef851edbc18"} Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.862456 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.862474 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.862526 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.864003 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.864020 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.864012 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.864051 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.864061 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.864033 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.864124 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.864036 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:35 crc kubenswrapper[4667]: I0929 17:09:35.864151 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:36 crc kubenswrapper[4667]: I0929 17:09:36.543878 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Sep 29 17:09:36 crc kubenswrapper[4667]: I0929 17:09:36.864083 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:36 crc kubenswrapper[4667]: I0929 17:09:36.864893 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:36 crc kubenswrapper[4667]: I0929 17:09:36.864930 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:36 crc kubenswrapper[4667]: I0929 17:09:36.864941 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.617757 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.866573 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.867284 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.867320 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.867331 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.957090 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.957164 4667 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.957191 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.957905 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.957933 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:37 crc kubenswrapper[4667]: I0929 17:09:37.957941 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:38 crc kubenswrapper[4667]: I0929 17:09:38.521740 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:38 crc kubenswrapper[4667]: I0929 17:09:38.868744 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:38 crc kubenswrapper[4667]: I0929 17:09:38.869414 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:38 crc kubenswrapper[4667]: I0929 17:09:38.869449 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:38 crc kubenswrapper[4667]: I0929 17:09:38.869460 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:41 crc kubenswrapper[4667]: E0929 17:09:41.873614 4667 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.662113 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.662254 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.663329 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.663459 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.663537 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.808553 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.876803 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.877567 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.877599 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:42 crc kubenswrapper[4667]: I0929 17:09:42.877608 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:43 crc kubenswrapper[4667]: I0929 17:09:43.624976 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:43 crc kubenswrapper[4667]: I0929 17:09:43.629562 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:43 crc kubenswrapper[4667]: I0929 17:09:43.784202 4667 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Sep 29 17:09:43 crc kubenswrapper[4667]: I0929 17:09:43.878282 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:43 crc kubenswrapper[4667]: I0929 17:09:43.878920 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:43 crc kubenswrapper[4667]: I0929 17:09:43.878948 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:43 crc kubenswrapper[4667]: I0929 17:09:43.878956 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:43 crc kubenswrapper[4667]: I0929 17:09:43.881247 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:44 crc kubenswrapper[4667]: I0929 17:09:44.382081 4667 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 29 17:09:44 crc kubenswrapper[4667]: I0929 17:09:44.382139 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Sep 29 17:09:44 crc kubenswrapper[4667]: I0929 17:09:44.431765 4667 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Sep 29 17:09:44 crc kubenswrapper[4667]: I0929 17:09:44.431830 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Sep 29 17:09:44 crc kubenswrapper[4667]: I0929 17:09:44.879811 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:44 crc kubenswrapper[4667]: I0929 17:09:44.880697 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:44 crc kubenswrapper[4667]: I0929 17:09:44.880737 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:44 crc kubenswrapper[4667]: I0929 17:09:44.880748 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:45 crc kubenswrapper[4667]: I0929 17:09:45.662192 4667 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Sep 29 17:09:45 crc kubenswrapper[4667]: I0929 17:09:45.662255 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Sep 29 17:09:45 crc kubenswrapper[4667]: I0929 17:09:45.882261 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:45 crc kubenswrapper[4667]: I0929 17:09:45.883003 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:45 crc kubenswrapper[4667]: I0929 17:09:45.883040 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:45 crc kubenswrapper[4667]: I0929 17:09:45.883051 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.635162 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.635305 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.636650 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.636689 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.636699 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.645470 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.886776 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.887573 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.887628 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:47 crc kubenswrapper[4667]: I0929 17:09:47.887682 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.384675 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.384804 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.385633 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.385669 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.385679 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.389633 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.420823 4667 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.421571 4667 trace.go:236] Trace[763028596]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Sep-2025 17:09:35.456) (total time: 13964ms): Sep 29 17:09:49 crc kubenswrapper[4667]: Trace[763028596]: ---"Objects listed" error: 13964ms (17:09:49.421) Sep 29 17:09:49 crc kubenswrapper[4667]: Trace[763028596]: [13.964932507s] [13.964932507s] END Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.421594 4667 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.422622 4667 trace.go:236] Trace[358797763]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Sep-2025 17:09:34.863) (total time: 14559ms): Sep 29 17:09:49 crc kubenswrapper[4667]: Trace[358797763]: ---"Objects listed" error: 14559ms (17:09:49.422) Sep 29 17:09:49 crc kubenswrapper[4667]: Trace[358797763]: [14.559068674s] [14.559068674s] END Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.422776 4667 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.423054 4667 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.423091 4667 trace.go:236] Trace[1029075385]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Sep-2025 17:09:35.204) (total time: 14218ms): Sep 29 17:09:49 crc kubenswrapper[4667]: Trace[1029075385]: ---"Objects listed" error: 14218ms (17:09:49.423) Sep 29 17:09:49 crc kubenswrapper[4667]: Trace[1029075385]: [14.218588552s] [14.218588552s] END Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.423105 4667 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.423351 4667 trace.go:236] Trace[142401484]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Sep-2025 17:09:35.485) (total time: 13938ms): Sep 29 17:09:49 crc kubenswrapper[4667]: Trace[142401484]: ---"Objects listed" error: 13938ms (17:09:49.423) Sep 29 17:09:49 crc kubenswrapper[4667]: Trace[142401484]: [13.938236445s] [13.938236445s] END Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.423376 4667 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.423637 4667 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.446358 4667 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43540->192.168.126.11:17697: read: connection reset by peer" start-of-body= Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.446404 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:43540->192.168.126.11:17697: read: connection reset by peer" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.446640 4667 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.446688 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.785188 4667 apiserver.go:52] "Watching apiserver" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.787118 4667 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.787332 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-lbbgf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.787872 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.787888 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.788033 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.788526 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lbbgf" Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.788604 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.788687 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.789024 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.789134 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.789230 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.789294 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.789470 4667 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.791788 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.791865 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.791796 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.792053 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.792219 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.793123 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.793159 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.793195 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.793258 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.793371 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.793399 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.793539 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.805273 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.815137 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.823279 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.825266 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.825981 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826034 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826054 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826069 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826084 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826097 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826111 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826124 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826138 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826151 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826164 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826184 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826198 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826212 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826235 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826248 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826266 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826280 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826294 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826308 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826320 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826335 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826350 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826371 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826384 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826397 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826410 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826425 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826440 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826454 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826469 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826481 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826494 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826507 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826520 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826534 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826548 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826563 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826591 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826610 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826633 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826659 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826674 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826705 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826721 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826735 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826749 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826764 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826780 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826796 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826813 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826827 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826855 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826869 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826883 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826909 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826933 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826946 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826941 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826960 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826975 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.826989 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827003 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827017 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827031 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827047 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827060 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827073 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827090 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827090 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827104 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827118 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827134 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827150 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827164 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827179 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827192 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827209 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827220 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827238 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827256 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827272 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827288 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827356 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827375 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827397 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827414 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827430 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827446 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827461 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827475 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827477 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827491 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827508 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827534 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827548 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827563 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827597 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827611 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827627 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827642 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827675 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827691 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827707 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827721 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827736 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827751 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827765 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827790 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827805 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827819 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827832 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827873 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827890 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827904 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827919 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827934 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827949 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827964 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827978 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.827992 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828008 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828010 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828022 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828038 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828054 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828067 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828082 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828097 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828112 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828126 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828144 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828683 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828723 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828875 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.829303 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.829376 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.829460 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.829524 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.829820 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.829898 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.829918 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.829916 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830000 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830015 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830357 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830576 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830687 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.830915 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:09:50.330891423 +0000 UTC m=+18.828738191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830923 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831029 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831114 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831120 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831238 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831254 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831257 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830763 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830770 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830793 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830734 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831327 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.830726 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831800 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831873 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.832011 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.832044 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.831736 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.832315 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.832385 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.832531 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.832542 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.832620 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.832864 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.832979 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833263 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833270 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833364 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833429 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833499 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833494 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833518 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833564 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833751 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833798 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833783 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834017 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.833961 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834278 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834132 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834400 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834429 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834433 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834535 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834712 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834783 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834907 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834966 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.834996 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.835001 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.835106 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.835163 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.835301 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.835321 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.835298 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.835430 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.835568 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.835449 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.828141 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836258 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836286 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836312 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836330 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836348 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836364 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836379 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836396 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836414 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836432 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836447 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836464 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836482 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836496 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836513 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836529 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836544 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836564 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836581 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836599 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836618 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836636 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836663 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836678 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836695 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836711 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836728 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836742 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836759 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836776 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836791 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836807 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836825 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836853 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836871 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836891 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836909 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836924 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836941 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836958 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836974 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.836991 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837008 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837022 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837038 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837056 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837073 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837089 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837105 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837122 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837138 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837157 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837176 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837195 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837210 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837228 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837246 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837262 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837279 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837295 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837309 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837326 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837343 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837359 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837374 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837392 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837409 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837425 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837443 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837490 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837509 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837531 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837549 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837573 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837588 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837608 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837629 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837658 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7192692e-a16e-46c5-9097-0e15418054a4-hosts-file\") pod \"node-resolver-lbbgf\" (UID: \"7192692e-a16e-46c5-9097-0e15418054a4\") " pod="openshift-dns/node-resolver-lbbgf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837676 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9hv6\" (UniqueName: \"kubernetes.io/projected/7192692e-a16e-46c5-9097-0e15418054a4-kube-api-access-q9hv6\") pod \"node-resolver-lbbgf\" (UID: \"7192692e-a16e-46c5-9097-0e15418054a4\") " pod="openshift-dns/node-resolver-lbbgf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837693 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837714 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837733 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837752 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837769 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837786 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837868 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837881 4667 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837891 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837900 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837909 4667 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837919 4667 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837928 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837936 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837945 4667 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837955 4667 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837963 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837973 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837983 4667 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837991 4667 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.837999 4667 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838007 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838020 4667 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838028 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838037 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838045 4667 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838055 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838064 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838074 4667 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838084 4667 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838094 4667 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838103 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838112 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838122 4667 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838131 4667 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838139 4667 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838146 4667 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838156 4667 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838164 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838173 4667 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838181 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838191 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838199 4667 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838208 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838219 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838227 4667 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838236 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838244 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838256 4667 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838264 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838272 4667 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838280 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838290 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838298 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838306 4667 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838316 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838325 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838334 4667 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838343 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838353 4667 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838361 4667 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838369 4667 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838378 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838387 4667 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838395 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838403 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838412 4667 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838422 4667 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838430 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838438 4667 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838448 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838456 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838464 4667 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838473 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838483 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838491 4667 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838499 4667 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838506 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838516 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838526 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838536 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838545 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838555 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838564 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838573 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838583 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838592 4667 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838599 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838715 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838828 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838864 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838939 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.838973 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839059 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839104 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839174 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839403 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839416 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839460 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839504 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839620 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839668 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839734 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839791 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.839878 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.841217 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.841515 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.841614 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.841789 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.842164 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.842170 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.842629 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.843115 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.843343 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.843509 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.843692 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.843898 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.843927 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.843960 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.843994 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.844054 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.844137 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.844461 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.844487 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.844629 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.848105 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.848310 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.848497 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.848689 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.848713 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.848932 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.849175 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.849258 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.849394 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.849561 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.849707 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.849908 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.850022 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.850288 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.850464 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.850607 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.850852 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.851272 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.851959 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.852436 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.852567 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.852637 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.852690 4667 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.853720 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.853918 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.853928 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.854517 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.855120 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.855219 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.855288 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.855419 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.855484 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.855688 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.855956 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.856242 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.856385 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.856472 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.856591 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.857307 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.857703 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.857937 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.858114 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.858310 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.858331 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.858496 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.858678 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.858745 4667 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.858809 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:50.358794637 +0000 UTC m=+18.856641406 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.858873 4667 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.858900 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.858915 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:50.358901046 +0000 UTC m=+18.856747815 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.859073 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.859112 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.859285 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.859302 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.859316 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.859004 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.859368 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.859561 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.859921 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.860053 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.860216 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.860240 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.860423 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.860441 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.860516 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.860771 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.860814 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.860946 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.861336 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.861339 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.861457 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.861827 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.861904 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.863750 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.864391 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.864952 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.865067 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.865504 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.865594 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.865957 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.879828 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.880191 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.881173 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.881474 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.881604 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.881638 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.881792 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.882597 4667 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.882672 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.882694 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.882708 4667 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.882754 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:50.382741291 +0000 UTC m=+18.880588060 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:49 crc kubenswrapper[4667]: E0929 17:09:49.882922 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:50.382909386 +0000 UTC m=+18.880756156 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.883620 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.890411 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.894926 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.898321 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.900112 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.901875 4667 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92" exitCode=255 Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.901915 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92"} Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.907540 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.909974 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.913865 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.914026 4667 scope.go:117] "RemoveContainer" containerID="e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.919867 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.926417 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.939516 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.939550 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7192692e-a16e-46c5-9097-0e15418054a4-hosts-file\") pod \"node-resolver-lbbgf\" (UID: \"7192692e-a16e-46c5-9097-0e15418054a4\") " pod="openshift-dns/node-resolver-lbbgf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.939570 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9hv6\" (UniqueName: \"kubernetes.io/projected/7192692e-a16e-46c5-9097-0e15418054a4-kube-api-access-q9hv6\") pod \"node-resolver-lbbgf\" (UID: \"7192692e-a16e-46c5-9097-0e15418054a4\") " pod="openshift-dns/node-resolver-lbbgf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.939566 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.939597 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.939689 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.939755 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7192692e-a16e-46c5-9097-0e15418054a4-hosts-file\") pod \"node-resolver-lbbgf\" (UID: \"7192692e-a16e-46c5-9097-0e15418054a4\") " pod="openshift-dns/node-resolver-lbbgf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940052 4667 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940134 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940165 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940183 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940214 4667 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940244 4667 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940265 4667 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940275 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940284 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940296 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940305 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940314 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940323 4667 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940330 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940338 4667 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940345 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940387 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940400 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940409 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940419 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940429 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940436 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940445 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940452 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940466 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940475 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940483 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940490 4667 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940507 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940537 4667 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940546 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940555 4667 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940563 4667 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940571 4667 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940579 4667 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940587 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940594 4667 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940602 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940610 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940618 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940626 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940634 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940642 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940659 4667 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940669 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940678 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940686 4667 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940694 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940702 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940709 4667 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940718 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940726 4667 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940733 4667 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940741 4667 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940749 4667 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940756 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940763 4667 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940771 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940778 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940787 4667 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940795 4667 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940802 4667 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940810 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940817 4667 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940825 4667 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940833 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940856 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940863 4667 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940870 4667 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940878 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940885 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940893 4667 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940901 4667 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940908 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940917 4667 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940926 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940935 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940943 4667 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940951 4667 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940958 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940967 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940974 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940981 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.940989 4667 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941010 4667 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941018 4667 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941025 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941031 4667 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941039 4667 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941046 4667 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941054 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941062 4667 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941069 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941077 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941084 4667 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941093 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941100 4667 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941107 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941115 4667 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941122 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941129 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941137 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941145 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941153 4667 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941161 4667 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941169 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941179 4667 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941187 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941194 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941202 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941210 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941229 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.941237 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.947337 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.952048 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9hv6\" (UniqueName: \"kubernetes.io/projected/7192692e-a16e-46c5-9097-0e15418054a4-kube-api-access-q9hv6\") pod \"node-resolver-lbbgf\" (UID: \"7192692e-a16e-46c5-9097-0e15418054a4\") " pod="openshift-dns/node-resolver-lbbgf" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.954216 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.960794 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.966826 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.971742 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:49 crc kubenswrapper[4667]: I0929 17:09:49.977678 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.098305 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.105911 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Sep 29 17:09:50 crc kubenswrapper[4667]: W0929 17:09:50.106819 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-adda0eece4944b99f6c561bd28f99c338f523d68c4cb70e2a667b18e17ee2c71 WatchSource:0}: Error finding container adda0eece4944b99f6c561bd28f99c338f523d68c4cb70e2a667b18e17ee2c71: Status 404 returned error can't find the container with id adda0eece4944b99f6c561bd28f99c338f523d68c4cb70e2a667b18e17ee2c71 Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.111985 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lbbgf" Sep 29 17:09:50 crc kubenswrapper[4667]: W0929 17:09:50.115975 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-f86de3d02d6c9e2bfbf259b3228a586eac248be7f273bc3080ab8f31f9d530a5 WatchSource:0}: Error finding container f86de3d02d6c9e2bfbf259b3228a586eac248be7f273bc3080ab8f31f9d530a5: Status 404 returned error can't find the container with id f86de3d02d6c9e2bfbf259b3228a586eac248be7f273bc3080ab8f31f9d530a5 Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.117135 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Sep 29 17:09:50 crc kubenswrapper[4667]: W0929 17:09:50.129563 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-de4376b3dd3d88c52225cfcdfcb3af9374b35f7bff9645013cee977c25caaa7a WatchSource:0}: Error finding container de4376b3dd3d88c52225cfcdfcb3af9374b35f7bff9645013cee977c25caaa7a: Status 404 returned error can't find the container with id de4376b3dd3d88c52225cfcdfcb3af9374b35f7bff9645013cee977c25caaa7a Sep 29 17:09:50 crc kubenswrapper[4667]: W0929 17:09:50.131896 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7192692e_a16e_46c5_9097_0e15418054a4.slice/crio-4952ab902ed9a1643255c9f3a199b202b0c5bb06b99e9568056e7a020e532eef WatchSource:0}: Error finding container 4952ab902ed9a1643255c9f3a199b202b0c5bb06b99e9568056e7a020e532eef: Status 404 returned error can't find the container with id 4952ab902ed9a1643255c9f3a199b202b0c5bb06b99e9568056e7a020e532eef Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.344628 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.344777 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:09:51.344756318 +0000 UTC m=+19.842603088 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.445173 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.445213 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.445235 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.445252 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445353 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445364 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445391 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445403 4667 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445369 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445446 4667 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445450 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:51.445437226 +0000 UTC m=+19.943283996 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445481 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:51.44547091 +0000 UTC m=+19.943317679 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445349 4667 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445353 4667 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445511 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:51.445504703 +0000 UTC m=+19.943351472 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:50 crc kubenswrapper[4667]: E0929 17:09:50.445566 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:51.445552392 +0000 UTC m=+19.943399162 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.905036 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.906357 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa"} Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.906530 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.907146 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lbbgf" event={"ID":"7192692e-a16e-46c5-9097-0e15418054a4","Type":"ContainerStarted","Data":"1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4"} Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.907179 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lbbgf" event={"ID":"7192692e-a16e-46c5-9097-0e15418054a4","Type":"ContainerStarted","Data":"4952ab902ed9a1643255c9f3a199b202b0c5bb06b99e9568056e7a020e532eef"} Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.908557 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4"} Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.908584 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e"} Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.908594 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f86de3d02d6c9e2bfbf259b3228a586eac248be7f273bc3080ab8f31f9d530a5"} Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.909539 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003"} Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.909568 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"adda0eece4944b99f6c561bd28f99c338f523d68c4cb70e2a667b18e17ee2c71"} Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.910520 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"de4376b3dd3d88c52225cfcdfcb3af9374b35f7bff9645013cee977c25caaa7a"} Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.916411 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.924238 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.932631 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.947555 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.969771 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:50 crc kubenswrapper[4667]: I0929 17:09:50.984240 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.004576 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:50Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.015020 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.024616 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.039321 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.053393 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.062083 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.068392 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.075720 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.083583 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.091780 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.242881 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-l8rmj"] Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.243192 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.245040 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-2bf6m"] Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.245316 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.247877 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.247877 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.249073 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.249129 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.249244 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.249369 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.249728 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.249754 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.249789 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.251487 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.258520 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.273170 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.286011 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.298012 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.306598 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.314607 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.322940 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.331852 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.338602 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.346687 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351308 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351364 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-os-release\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351392 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/28fa0016-3e75-4704-8b60-30ee9e576d59-rootfs\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351410 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh4nc\" (UniqueName: \"kubernetes.io/projected/28fa0016-3e75-4704-8b60-30ee9e576d59-kube-api-access-wh4nc\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351423 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-hostroot\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.351464 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:09:53.351440933 +0000 UTC m=+21.849287702 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351501 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-conf-dir\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351574 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28fa0016-3e75-4704-8b60-30ee9e576d59-proxy-tls\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351620 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-socket-dir-parent\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351651 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-var-lib-kubelet\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351667 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-cni-dir\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351680 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/788f770a-3181-4b66-981c-90ffb7fc49c0-cni-binary-copy\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351703 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28fa0016-3e75-4704-8b60-30ee9e576d59-mcd-auth-proxy-config\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351716 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-run-netns\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351728 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-var-lib-cni-bin\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351741 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-run-multus-certs\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351753 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-etc-kubernetes\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351767 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-system-cni-dir\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351783 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-cnibin\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351795 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-run-k8s-cni-cncf-io\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351809 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q54hx\" (UniqueName: \"kubernetes.io/projected/788f770a-3181-4b66-981c-90ffb7fc49c0-kube-api-access-q54hx\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351826 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-daemon-config\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.351855 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-var-lib-cni-multus\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.355190 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.363048 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.369968 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.380674 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.395048 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.415811 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.429879 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.438041 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.448008 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452356 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452392 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-cnibin\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452418 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-run-k8s-cni-cncf-io\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452442 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q54hx\" (UniqueName: \"kubernetes.io/projected/788f770a-3181-4b66-981c-90ffb7fc49c0-kube-api-access-q54hx\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.452450 4667 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452505 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-cnibin\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452515 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-var-lib-cni-multus\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.452519 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:53.452502586 +0000 UTC m=+21.950349355 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452463 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-var-lib-cni-multus\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452587 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-run-k8s-cni-cncf-io\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452604 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-daemon-config\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452669 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-os-release\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452702 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452720 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/28fa0016-3e75-4704-8b60-30ee9e576d59-rootfs\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452744 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh4nc\" (UniqueName: \"kubernetes.io/projected/28fa0016-3e75-4704-8b60-30ee9e576d59-kube-api-access-wh4nc\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452760 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-hostroot\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452773 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-conf-dir\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452799 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28fa0016-3e75-4704-8b60-30ee9e576d59-proxy-tls\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452813 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-socket-dir-parent\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452818 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-conf-dir\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452852 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.452855 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452869 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-var-lib-kubelet\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.452874 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.452889 4667 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452887 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-cni-dir\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452905 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/788f770a-3181-4b66-981c-90ffb7fc49c0-cni-binary-copy\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452895 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-hostroot\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.452928 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:53.452915201 +0000 UTC m=+21.950761971 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452934 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-var-lib-kubelet\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452911 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-socket-dir-parent\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452816 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/28fa0016-3e75-4704-8b60-30ee9e576d59-rootfs\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.452937 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452948 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28fa0016-3e75-4704-8b60-30ee9e576d59-mcd-auth-proxy-config\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.453003 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453014 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.453021 4667 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453034 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-system-cni-dir\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.452908 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-os-release\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.453057 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:53.453046668 +0000 UTC m=+21.950893437 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453036 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-cni-dir\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453076 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-system-cni-dir\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453089 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-run-netns\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453080 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-run-netns\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.453096 4667 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453118 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-var-lib-cni-bin\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453142 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-var-lib-cni-bin\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.453146 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:53.453129974 +0000 UTC m=+21.950976743 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453169 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-run-multus-certs\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453184 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-host-run-multus-certs\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453192 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-etc-kubernetes\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453227 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/788f770a-3181-4b66-981c-90ffb7fc49c0-etc-kubernetes\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453199 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/788f770a-3181-4b66-981c-90ffb7fc49c0-multus-daemon-config\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453433 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/788f770a-3181-4b66-981c-90ffb7fc49c0-cni-binary-copy\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.453474 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/28fa0016-3e75-4704-8b60-30ee9e576d59-mcd-auth-proxy-config\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.456624 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/28fa0016-3e75-4704-8b60-30ee9e576d59-proxy-tls\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.467567 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q54hx\" (UniqueName: \"kubernetes.io/projected/788f770a-3181-4b66-981c-90ffb7fc49c0-kube-api-access-q54hx\") pod \"multus-2bf6m\" (UID: \"788f770a-3181-4b66-981c-90ffb7fc49c0\") " pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.469036 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh4nc\" (UniqueName: \"kubernetes.io/projected/28fa0016-3e75-4704-8b60-30ee9e576d59-kube-api-access-wh4nc\") pod \"machine-config-daemon-l8rmj\" (UID: \"28fa0016-3e75-4704-8b60-30ee9e576d59\") " pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.554919 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.559724 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2bf6m" Sep 29 17:09:51 crc kubenswrapper[4667]: W0929 17:09:51.563131 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28fa0016_3e75_4704_8b60_30ee9e576d59.slice/crio-331825814b1ac694c2e820924659cf3fe276388a9c27e68b6d9606f6cffbbbad WatchSource:0}: Error finding container 331825814b1ac694c2e820924659cf3fe276388a9c27e68b6d9606f6cffbbbad: Status 404 returned error can't find the container with id 331825814b1ac694c2e820924659cf3fe276388a9c27e68b6d9606f6cffbbbad Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.615799 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qjsnt"] Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.616700 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-5mghc"] Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.616879 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.617377 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.619243 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.619653 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.619682 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.620712 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.620720 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.620887 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.620923 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.620937 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.622983 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.632409 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.642045 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.654037 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.664767 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.674266 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.684688 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.693299 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.706474 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.718730 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.727157 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.735029 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.743437 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.752863 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756643 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-netns\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756674 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756693 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-var-lib-openvswitch\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756710 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-cni-binary-copy\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756726 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-systemd\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756775 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-etc-openvswitch\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756806 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqs9s\" (UniqueName: \"kubernetes.io/projected/6250a9ff-80f5-44d8-90f6-40e77106af6c-kube-api-access-bqs9s\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756867 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-script-lib\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756896 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhgkq\" (UniqueName: \"kubernetes.io/projected/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-kube-api-access-dhgkq\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756915 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-node-log\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756946 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-system-cni-dir\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756961 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-os-release\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756974 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-kubelet\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756986 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-systemd-units\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.756999 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-netd\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757014 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757027 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-openvswitch\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757040 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-ovn\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757053 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-bin\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757076 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-cnibin\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757089 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-env-overrides\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757102 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovn-node-metrics-cert\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757119 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757134 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-slash\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757150 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-ovn-kubernetes\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757173 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-config\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.757187 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-log-socket\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.763354 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.771714 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.780931 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.790103 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.800041 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.809439 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.815103 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.815127 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.815225 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.815277 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.815330 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:09:51 crc kubenswrapper[4667]: E0929 17:09:51.815438 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.818601 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.818816 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.819232 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.820094 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.821120 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.821708 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.822655 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.823386 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.824283 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.824926 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.825408 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.826259 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.826917 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.827720 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.827924 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.828218 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.829060 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.829536 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.830435 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.830981 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.831486 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.832387 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.832808 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.833361 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.834169 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.834763 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.835556 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.836190 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.837601 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.838237 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.839146 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.839697 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.840227 4667 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.840340 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.841806 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.841788 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.842417 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.842921 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.844340 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.847927 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.848654 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.849901 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.852050 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.852520 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.853487 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.854447 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.855027 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.855811 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.856416 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.856705 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857660 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-script-lib\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857701 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhgkq\" (UniqueName: \"kubernetes.io/projected/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-kube-api-access-dhgkq\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857728 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-node-log\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857746 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-kubelet\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857770 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-system-cni-dir\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857787 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-os-release\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857802 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-systemd-units\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857818 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-netd\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857836 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857874 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-openvswitch\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857892 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-ovn\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857909 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-bin\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857931 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-system-cni-dir\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.857942 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovn-node-metrics-cert\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858021 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-cnibin\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858043 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-env-overrides\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858060 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-ovn-kubernetes\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858087 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858105 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-slash\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858125 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-cnibin\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858132 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-config\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858178 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-log-socket\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858207 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-netns\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858225 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858247 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-var-lib-openvswitch\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858265 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-cni-binary-copy\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858291 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-systemd\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858336 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-etc-openvswitch\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858351 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqs9s\" (UniqueName: \"kubernetes.io/projected/6250a9ff-80f5-44d8-90f6-40e77106af6c-kube-api-access-bqs9s\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858355 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-script-lib\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858412 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858414 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-ovn-kubernetes\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858462 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-kubelet\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858484 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-log-socket\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858504 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-netns\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858546 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-var-lib-openvswitch\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858565 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858646 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-config\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858687 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-slash\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858714 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-systemd\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858739 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-ovn\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858762 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-openvswitch\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858786 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-systemd-units\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858802 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-env-overrides\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858870 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-netd\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858898 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-etc-openvswitch\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858883 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-node-log\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858912 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-os-release\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.858941 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-bin\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.859118 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-cni-binary-copy\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.859118 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.859160 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.859868 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.860331 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.861123 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.861564 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.862384 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.862974 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.863308 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovn-node-metrics-cert\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.863408 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.865833 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.873111 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqs9s\" (UniqueName: \"kubernetes.io/projected/6250a9ff-80f5-44d8-90f6-40e77106af6c-kube-api-access-bqs9s\") pod \"ovnkube-node-qjsnt\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.873326 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhgkq\" (UniqueName: \"kubernetes.io/projected/61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4-kube-api-access-dhgkq\") pod \"multus-additional-cni-plugins-5mghc\" (UID: \"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\") " pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.876933 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.887322 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.897002 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.906464 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.914542 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860"} Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.914575 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657"} Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.914586 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"331825814b1ac694c2e820924659cf3fe276388a9c27e68b6d9606f6cffbbbad"} Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.915690 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6"} Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.919779 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2bf6m" event={"ID":"788f770a-3181-4b66-981c-90ffb7fc49c0","Type":"ContainerStarted","Data":"1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c"} Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.919820 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2bf6m" event={"ID":"788f770a-3181-4b66-981c-90ffb7fc49c0","Type":"ContainerStarted","Data":"7672c25cf6a9dc8eda1c6628685fcddf5f254762be1795b19836314e0f2e01eb"} Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.926325 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.932094 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5mghc" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.935814 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:51 crc kubenswrapper[4667]: I0929 17:09:51.975731 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.014926 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.056472 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.095172 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.137475 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.179543 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.217949 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.255251 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.295752 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.334246 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.374560 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.414076 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.455608 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.493229 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.534752 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.573740 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.613607 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.623736 4667 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.625833 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.625883 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.625892 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.625982 4667 kubelet_node_status.go:76] "Attempting to register node" node="crc" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.665445 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.668151 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.677637 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.688250 4667 kubelet_node_status.go:115] "Node was previously registered" node="crc" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.688475 4667 kubelet_node_status.go:79] "Successfully registered node" node="crc" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.689754 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.689785 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.689795 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.689808 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.689818 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:52Z","lastTransitionTime":"2025-09-29T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:52 crc kubenswrapper[4667]: E0929 17:09:52.704166 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.706499 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.706526 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.706535 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.706547 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.706556 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:52Z","lastTransitionTime":"2025-09-29T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.712324 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Sep 29 17:09:52 crc kubenswrapper[4667]: E0929 17:09:52.714500 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.716692 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.716722 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.716731 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.716744 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.716752 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:52Z","lastTransitionTime":"2025-09-29T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:52 crc kubenswrapper[4667]: E0929 17:09:52.725017 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.727270 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.727296 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.727305 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.727317 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.727326 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:52Z","lastTransitionTime":"2025-09-29T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:52 crc kubenswrapper[4667]: E0929 17:09:52.735103 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.737549 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.737579 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.737588 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.737600 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.737607 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:52Z","lastTransitionTime":"2025-09-29T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:52 crc kubenswrapper[4667]: E0929 17:09:52.748637 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: E0929 17:09:52.748753 4667 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.750023 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.750049 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.750059 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.750070 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.750077 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:52Z","lastTransitionTime":"2025-09-29T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.754126 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.798378 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.833874 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.852731 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.852765 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.852773 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.852786 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.852799 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:52Z","lastTransitionTime":"2025-09-29T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.872188 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.913803 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.924424 4667 generic.go:334] "Generic (PLEG): container finished" podID="61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4" containerID="8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e" exitCode=0 Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.924489 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" event={"ID":"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4","Type":"ContainerDied","Data":"8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e"} Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.924516 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" event={"ID":"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4","Type":"ContainerStarted","Data":"5367adc073ef65f86d85fd48d9c16e05bd1846f1e1b1a521ba9037041a524dfc"} Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.925611 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a" exitCode=0 Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.925681 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a"} Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.925722 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"87837f7510257125171781c62deb4919c2d91e84936b76106ade551402416775"} Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.954296 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:52Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.954663 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.954695 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.954703 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.954716 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:52 crc kubenswrapper[4667]: I0929 17:09:52.954726 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:52Z","lastTransitionTime":"2025-09-29T17:09:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.002280 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.034353 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.057987 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.058014 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.058022 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.058035 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.058043 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.073820 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.115303 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.158205 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.161145 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.161187 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.161208 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.161228 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.161239 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.195278 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.235320 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.262828 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.262872 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.262880 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.262895 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.262904 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.277273 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.313657 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.356276 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.364526 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.364556 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.364565 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.364579 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.364589 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.371963 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.372128 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:09:57.37210552 +0000 UTC m=+25.869952299 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.394548 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.415740 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-h7cv4"] Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.416143 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.434285 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.448630 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.466910 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.466947 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.466957 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.466972 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.466981 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.467541 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.472759 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.472798 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6g4c\" (UniqueName: \"kubernetes.io/projected/812bd13c-b2d7-4e1a-a226-5794831f8c6a-kube-api-access-n6g4c\") pod \"node-ca-h7cv4\" (UID: \"812bd13c-b2d7-4e1a-a226-5794831f8c6a\") " pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.472821 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/812bd13c-b2d7-4e1a-a226-5794831f8c6a-host\") pod \"node-ca-h7cv4\" (UID: \"812bd13c-b2d7-4e1a-a226-5794831f8c6a\") " pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.472857 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.472879 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.472890 4667 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.472929 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.472948 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:57.472932322 +0000 UTC m=+25.970779101 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.472970 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.472975 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/812bd13c-b2d7-4e1a-a226-5794831f8c6a-serviceca\") pod \"node-ca-h7cv4\" (UID: \"812bd13c-b2d7-4e1a-a226-5794831f8c6a\") " pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.472984 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.472996 4667 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.473007 4667 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.473029 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:57.473018014 +0000 UTC m=+25.970864783 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.473047 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:57.473035356 +0000 UTC m=+25.970882135 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.473093 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.473127 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.473141 4667 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.473199 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 17:09:57.473179787 +0000 UTC m=+25.971026566 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.488245 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.508066 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.554306 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.569564 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.569595 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.569604 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.569631 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.569640 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.573823 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6g4c\" (UniqueName: \"kubernetes.io/projected/812bd13c-b2d7-4e1a-a226-5794831f8c6a-kube-api-access-n6g4c\") pod \"node-ca-h7cv4\" (UID: \"812bd13c-b2d7-4e1a-a226-5794831f8c6a\") " pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.573872 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/812bd13c-b2d7-4e1a-a226-5794831f8c6a-host\") pod \"node-ca-h7cv4\" (UID: \"812bd13c-b2d7-4e1a-a226-5794831f8c6a\") " pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.573927 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/812bd13c-b2d7-4e1a-a226-5794831f8c6a-serviceca\") pod \"node-ca-h7cv4\" (UID: \"812bd13c-b2d7-4e1a-a226-5794831f8c6a\") " pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.573989 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/812bd13c-b2d7-4e1a-a226-5794831f8c6a-host\") pod \"node-ca-h7cv4\" (UID: \"812bd13c-b2d7-4e1a-a226-5794831f8c6a\") " pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.574673 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/812bd13c-b2d7-4e1a-a226-5794831f8c6a-serviceca\") pod \"node-ca-h7cv4\" (UID: \"812bd13c-b2d7-4e1a-a226-5794831f8c6a\") " pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.593091 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.621727 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6g4c\" (UniqueName: \"kubernetes.io/projected/812bd13c-b2d7-4e1a-a226-5794831f8c6a-kube-api-access-n6g4c\") pod \"node-ca-h7cv4\" (UID: \"812bd13c-b2d7-4e1a-a226-5794831f8c6a\") " pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.652685 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.672179 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.672214 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.672223 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.672238 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.672247 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.699875 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.725017 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-h7cv4" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.733170 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: W0929 17:09:53.735595 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod812bd13c_b2d7_4e1a_a226_5794831f8c6a.slice/crio-a78ef4189912e0424bddeb35322fdeec6426278e3fae3b662819067abe1df2e1 WatchSource:0}: Error finding container a78ef4189912e0424bddeb35322fdeec6426278e3fae3b662819067abe1df2e1: Status 404 returned error can't find the container with id a78ef4189912e0424bddeb35322fdeec6426278e3fae3b662819067abe1df2e1 Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.774216 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.774267 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.774277 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.774335 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.774362 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.774739 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.814128 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.815187 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.815230 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.815296 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.815293 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.815383 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:09:53 crc kubenswrapper[4667]: E0929 17:09:53.815457 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.852511 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.876136 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.876171 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.876181 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.876194 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.876203 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.893814 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.930070 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-h7cv4" event={"ID":"812bd13c-b2d7-4e1a-a226-5794831f8c6a","Type":"ContainerStarted","Data":"1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.930108 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-h7cv4" event={"ID":"812bd13c-b2d7-4e1a-a226-5794831f8c6a","Type":"ContainerStarted","Data":"a78ef4189912e0424bddeb35322fdeec6426278e3fae3b662819067abe1df2e1"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.933091 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.933117 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.933126 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.933135 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.933143 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.933150 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.934465 4667 generic.go:334] "Generic (PLEG): container finished" podID="61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4" containerID="3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7" exitCode=0 Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.934505 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" event={"ID":"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4","Type":"ContainerDied","Data":"3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7"} Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.936523 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.973327 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.978573 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.978599 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.978607 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.978635 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:53 crc kubenswrapper[4667]: I0929 17:09:53.978645 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:53Z","lastTransitionTime":"2025-09-29T17:09:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.014872 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.054690 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.079835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.079879 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.079890 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.079905 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.079914 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.094454 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.135021 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.177730 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.182100 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.182124 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.182133 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.182148 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.182158 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.213971 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.252929 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.284046 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.284080 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.284088 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.284102 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.284110 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.293743 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.333051 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.375285 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.385583 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.385623 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.385635 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.385648 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.385656 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.413104 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.459752 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.487390 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.487426 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.487436 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.487449 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.487457 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.496564 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.534924 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.574755 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.589395 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.589429 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.589439 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.589453 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.589461 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.614228 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.655276 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.691414 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.691443 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.691451 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.691464 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.691473 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.694126 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.735681 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.774556 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.792493 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.792518 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.792527 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.792539 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.792547 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.813186 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.853707 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.894640 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.894671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.894681 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.894694 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.894703 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.897253 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.934162 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.938147 4667 generic.go:334] "Generic (PLEG): container finished" podID="61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4" containerID="e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5" exitCode=0 Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.938177 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" event={"ID":"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4","Type":"ContainerDied","Data":"e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5"} Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.974515 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.996358 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.996382 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.996390 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.996404 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:54 crc kubenswrapper[4667]: I0929 17:09:54.996412 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:54Z","lastTransitionTime":"2025-09-29T17:09:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.014459 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.053985 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.094150 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.098166 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.098192 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.098201 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.098214 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.098224 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:55Z","lastTransitionTime":"2025-09-29T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.134896 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.173036 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.199788 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.199813 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.199823 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.199835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.199861 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:55Z","lastTransitionTime":"2025-09-29T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.212132 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.253485 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.294246 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.301542 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.301566 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.301573 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.301584 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.301592 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:55Z","lastTransitionTime":"2025-09-29T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.334303 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.375615 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.402888 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.402921 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.402931 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.402942 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.402951 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:55Z","lastTransitionTime":"2025-09-29T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.414178 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.458171 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.494115 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.504431 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.504451 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.504461 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.504474 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.504482 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:55Z","lastTransitionTime":"2025-09-29T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.606521 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.606678 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.606741 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.606797 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.606860 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:55Z","lastTransitionTime":"2025-09-29T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.708749 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.708782 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.708793 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.708809 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.708819 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:55Z","lastTransitionTime":"2025-09-29T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.810254 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.810281 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.810291 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.810305 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.810312 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:55Z","lastTransitionTime":"2025-09-29T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.815657 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:55 crc kubenswrapper[4667]: E0929 17:09:55.815744 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.815877 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:55 crc kubenswrapper[4667]: E0929 17:09:55.815990 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.816025 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:55 crc kubenswrapper[4667]: E0929 17:09:55.816165 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.911975 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.912009 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.912018 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.912031 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.912039 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:55Z","lastTransitionTime":"2025-09-29T17:09:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.942434 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.943906 4667 generic.go:334] "Generic (PLEG): container finished" podID="61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4" containerID="ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390" exitCode=0 Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.943930 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" event={"ID":"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4","Type":"ContainerDied","Data":"ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390"} Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.953697 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.971855 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.984011 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.991676 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:55 crc kubenswrapper[4667]: I0929 17:09:55.999251 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:55Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.006440 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.014205 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.014233 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.014242 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.014254 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.014262 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.015103 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.026872 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.037317 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.045611 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.052420 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.060192 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.070539 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.079365 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.117007 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.117040 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.117054 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.117072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.117082 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.218815 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.218862 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.218873 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.218887 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.218895 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.320671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.320701 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.320711 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.320724 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.320732 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.421990 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.422021 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.422030 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.422044 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.422053 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.524101 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.524124 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.524131 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.524142 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.524150 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.625678 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.625702 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.625709 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.625719 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.625727 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.727877 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.727914 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.727922 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.727937 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.727947 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.829897 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.829932 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.829943 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.829960 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.829971 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.932199 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.932231 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.932242 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.932254 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.932263 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:56Z","lastTransitionTime":"2025-09-29T17:09:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.948853 4667 generic.go:334] "Generic (PLEG): container finished" podID="61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4" containerID="3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e" exitCode=0 Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.948868 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" event={"ID":"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4","Type":"ContainerDied","Data":"3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e"} Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.960664 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.969936 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.977901 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.987634 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:56 crc kubenswrapper[4667]: I0929 17:09:56.998629 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:56Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.011552 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.019853 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.027336 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.033675 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.033893 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.033909 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.033922 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.033932 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.035044 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.041798 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.049851 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.061768 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.070001 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.078323 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.135448 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.135477 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.135486 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.135499 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.135508 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.237347 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.237377 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.237386 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.237398 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.237406 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.339414 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.339440 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.339449 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.339459 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.339466 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.407028 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.407184 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:10:05.407172476 +0000 UTC m=+33.905019245 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.441310 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.441336 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.441344 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.441354 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.441361 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.507489 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.507527 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.507548 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.507566 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507610 4667 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507650 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:05.507640155 +0000 UTC m=+34.005486923 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507653 4667 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507659 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507675 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507685 4667 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507696 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:05.507684719 +0000 UTC m=+34.005531487 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507715 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:05.50770682 +0000 UTC m=+34.005553589 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507747 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507758 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507767 4667 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.507789 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:05.507783714 +0000 UTC m=+34.005630483 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.542767 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.542793 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.542802 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.542814 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.542824 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.645025 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.645063 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.645072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.645085 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.645093 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.746739 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.746787 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.746795 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.746809 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.746817 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.815189 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.815212 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.815239 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.815293 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.815377 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:09:57 crc kubenswrapper[4667]: E0929 17:09:57.815433 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.848616 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.848648 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.848658 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.848677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.848685 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.950953 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.950984 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.950992 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.951004 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.951013 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:57Z","lastTransitionTime":"2025-09-29T17:09:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.956161 4667 generic.go:334] "Generic (PLEG): container finished" podID="61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4" containerID="bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6" exitCode=0 Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.956214 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" event={"ID":"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4","Type":"ContainerDied","Data":"bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.960096 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940"} Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.960317 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.967226 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.977300 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.978301 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.989068 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:57 crc kubenswrapper[4667]: I0929 17:09:57.997935 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:57Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.004315 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.011431 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.019010 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.031372 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.038992 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.046939 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.053416 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.053602 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.053638 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.053648 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.053662 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.053671 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.061598 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.071070 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.079262 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.086063 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.095967 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.103778 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.111922 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.120236 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.128350 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.135883 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.142446 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.149762 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.155339 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.155367 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.155376 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.155387 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.155396 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.158829 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.165447 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.174534 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.186987 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.194741 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.256646 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.256769 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.256836 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.256932 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.256993 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.359210 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.359238 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.359246 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.359258 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.359280 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.460904 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.460939 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.460947 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.460959 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.460968 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.563303 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.563334 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.563342 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.563352 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.563359 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.664762 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.664788 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.664795 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.664805 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.664812 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.766831 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.766876 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.766885 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.766896 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.766904 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.868241 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.868275 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.868283 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.868295 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.868304 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.968989 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" event={"ID":"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4","Type":"ContainerStarted","Data":"f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.969032 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.969007 4667 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.969130 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.969146 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.969154 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.969165 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.969173 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:58Z","lastTransitionTime":"2025-09-29T17:09:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.978998 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.986830 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:09:58 crc kubenswrapper[4667]: I0929 17:09:58.991392 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:58Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.007337 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.014964 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.023867 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.033310 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.041326 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.049579 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.057050 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.063744 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.071017 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.071062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.071072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.071083 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.071091 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.071931 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.083592 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.091604 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.097693 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.106963 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.114889 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.122619 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.130614 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.138366 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.145883 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.153003 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.161803 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.172815 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.172866 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.172876 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.172888 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.172897 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.174090 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.182660 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.188915 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.197115 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.204768 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.214100 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.274397 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.274426 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.274435 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.274450 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.274458 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.376144 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.376174 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.376184 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.376196 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.376204 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.477969 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.478002 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.478010 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.478022 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.478030 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.579717 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.579750 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.579758 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.579770 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.579778 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.681798 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.681827 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.681863 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.681875 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.681883 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.783581 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.783616 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.783625 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.783638 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.783648 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.815128 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.815157 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.815157 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:09:59 crc kubenswrapper[4667]: E0929 17:09:59.815230 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:09:59 crc kubenswrapper[4667]: E0929 17:09:59.815301 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:09:59 crc kubenswrapper[4667]: E0929 17:09:59.815372 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.885291 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.885326 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.885336 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.885350 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.885361 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.974330 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/0.log" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.976239 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940" exitCode=1 Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.976260 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.976751 4667 scope.go:117] "RemoveContainer" containerID="f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.983920 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.986806 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.986832 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.986851 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.986863 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.986871 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:09:59Z","lastTransitionTime":"2025-09-29T17:09:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:09:59 crc kubenswrapper[4667]: I0929 17:09:59.991963 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:09:59Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.001298 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.010271 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.020177 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.028904 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.037043 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.044602 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.051868 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.059062 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.070467 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:09:59Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 17:09:59.431520 5952 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 17:09:59.431571 5952 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.431616 5952 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.431663 5952 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.431696 5952 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 17:09:59.431912 5952 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 17:09:59.432193 5952 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.432481 5952 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.077551 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.086230 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.088664 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.088701 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.088710 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.088722 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.088730 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:00Z","lastTransitionTime":"2025-09-29T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.095171 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.190892 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.191103 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.191111 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.191124 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.191133 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:00Z","lastTransitionTime":"2025-09-29T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.293339 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.293377 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.293387 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.293402 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.293412 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:00Z","lastTransitionTime":"2025-09-29T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.395588 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.395621 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.395630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.395642 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.395651 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:00Z","lastTransitionTime":"2025-09-29T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.496959 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.496985 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.496993 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.497005 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.497015 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:00Z","lastTransitionTime":"2025-09-29T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.599127 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.599161 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.599169 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.599182 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.599190 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:00Z","lastTransitionTime":"2025-09-29T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.702565 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.702602 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.702613 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.702629 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.702642 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:00Z","lastTransitionTime":"2025-09-29T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.805002 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.805032 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.805041 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.805054 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.805063 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:00Z","lastTransitionTime":"2025-09-29T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.906819 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.907088 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.907159 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.907243 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.907306 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:00Z","lastTransitionTime":"2025-09-29T17:10:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.979585 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/1.log" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.980083 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/0.log" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.981990 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05" exitCode=1 Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.982020 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05"} Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.982053 4667 scope.go:117] "RemoveContainer" containerID="f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.982496 4667 scope.go:117] "RemoveContainer" containerID="b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05" Sep 29 17:10:00 crc kubenswrapper[4667]: E0929 17:10:00.982634 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.991209 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:00 crc kubenswrapper[4667]: I0929 17:10:00.998869 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.008450 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.009113 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.009140 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.009149 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.009160 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.009169 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.016521 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.024690 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.031338 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.038586 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.046989 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.055106 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.061871 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.069674 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.082415 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:09:59Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 17:09:59.431520 5952 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 17:09:59.431571 5952 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.431616 5952 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.431663 5952 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.431696 5952 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 17:09:59.431912 5952 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 17:09:59.432193 5952 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.432481 5952 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.090183 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.096422 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.111056 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.111077 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.111086 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.111097 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.111105 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.213159 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.213186 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.213195 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.213207 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.213215 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.314858 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.314891 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.314900 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.314909 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.314917 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.416334 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.416363 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.416372 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.416383 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.416394 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.517926 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.517958 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.517968 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.517980 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.517991 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.620328 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.620366 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.620376 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.620389 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.620400 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.722570 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.722615 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.722624 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.722640 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.722650 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.815137 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:01 crc kubenswrapper[4667]: E0929 17:10:01.815243 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.815288 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.815324 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:01 crc kubenswrapper[4667]: E0929 17:10:01.815416 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:01 crc kubenswrapper[4667]: E0929 17:10:01.815333 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.823249 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.824020 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.824124 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.824181 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.824241 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.824292 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.831441 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.839085 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.847933 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.854817 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.863164 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.872789 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.881598 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.890619 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.896960 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.904559 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.913342 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.925277 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f47a5ed0196875373121a8dd165585fe15d5e53a08987771a55175de19901940\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:09:59Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 17:09:59.431520 5952 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0929 17:09:59.431571 5952 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.431616 5952 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.431663 5952 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.431696 5952 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 17:09:59.431912 5952 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0929 17:09:59.432193 5952 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0929 17:09:59.432481 5952 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.925522 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.925563 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.925574 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.925587 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.925598 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:01Z","lastTransitionTime":"2025-09-29T17:10:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.933998 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.985079 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/1.log" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.987320 4667 scope.go:117] "RemoveContainer" containerID="b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05" Sep 29 17:10:01 crc kubenswrapper[4667]: E0929 17:10:01.987446 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:10:01 crc kubenswrapper[4667]: I0929 17:10:01.995964 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.003160 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.012732 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.021127 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.027329 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.027354 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.027362 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.027372 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.027380 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.028959 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.037559 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.045656 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.053564 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.060287 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.067163 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.076351 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.083680 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.091637 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.103910 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.128971 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.129003 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.129012 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.129024 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.129033 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.230934 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.230960 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.230969 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.230979 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.230987 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.332867 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.332928 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.332939 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.332954 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.332968 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.434818 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.434875 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.434885 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.434898 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.434907 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.536612 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.536640 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.536649 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.536662 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.536675 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.592113 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29"] Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.592478 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.593891 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.594336 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.602381 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.610016 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.618187 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.626396 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.633668 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.638951 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.638986 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.638996 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.639009 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.639016 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.642055 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.648350 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.655345 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.663712 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.670741 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.677747 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.685500 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.697372 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.704857 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.711548 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.741034 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.741060 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.741069 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.741081 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.741090 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.745448 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmmr8\" (UniqueName: \"kubernetes.io/projected/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-kube-api-access-jmmr8\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.745476 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.745502 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.745522 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.767868 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.767971 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.768035 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.768091 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.768159 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: E0929 17:10:02.775694 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.777951 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.778124 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.778195 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.778256 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.778310 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: E0929 17:10:02.785896 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.788044 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.788073 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.788083 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.788094 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.788102 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: E0929 17:10:02.795603 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.797680 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.797707 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.797717 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.797727 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.797735 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: E0929 17:10:02.805666 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.807403 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.807423 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.807431 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.807439 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.807446 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: E0929 17:10:02.815483 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:02Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:02 crc kubenswrapper[4667]: E0929 17:10:02.815728 4667 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.842872 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.843023 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.843081 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.843143 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.843192 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.846252 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmmr8\" (UniqueName: \"kubernetes.io/projected/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-kube-api-access-jmmr8\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.846283 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.846310 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.846331 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.846993 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.847132 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.851392 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.858880 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmmr8\" (UniqueName: \"kubernetes.io/projected/64e4a86f-c3e1-4a00-b3c4-1d1294635c9a-kube-api-access-jmmr8\") pod \"ovnkube-control-plane-749d76644c-ljv29\" (UID: \"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.901640 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" Sep 29 17:10:02 crc kubenswrapper[4667]: W0929 17:10:02.910486 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64e4a86f_c3e1_4a00_b3c4_1d1294635c9a.slice/crio-a1fa2a18f9c3536c4aa78ab90704ae5d6897f9af5a554d75caa310051fe0ee9a WatchSource:0}: Error finding container a1fa2a18f9c3536c4aa78ab90704ae5d6897f9af5a554d75caa310051fe0ee9a: Status 404 returned error can't find the container with id a1fa2a18f9c3536c4aa78ab90704ae5d6897f9af5a554d75caa310051fe0ee9a Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.945755 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.945787 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.945796 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.945808 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.945817 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:02Z","lastTransitionTime":"2025-09-29T17:10:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:02 crc kubenswrapper[4667]: I0929 17:10:02.990994 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" event={"ID":"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a","Type":"ContainerStarted","Data":"a1fa2a18f9c3536c4aa78ab90704ae5d6897f9af5a554d75caa310051fe0ee9a"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.047185 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.047217 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.047227 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.047240 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.047249 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.148774 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.148807 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.148816 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.148828 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.148837 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.250799 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.250831 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.250853 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.250867 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.250876 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.352816 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.352861 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.352870 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.352883 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.352892 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.454298 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.454326 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.454336 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.454348 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.454357 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.556043 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.556074 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.556082 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.556094 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.556104 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.607824 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-cl5p9"] Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.608196 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:03 crc kubenswrapper[4667]: E0929 17:10:03.608252 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.616472 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.623253 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.629738 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.637717 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.645113 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.654211 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.657575 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.657601 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.657610 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.657621 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.657628 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.661426 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.669685 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.677596 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.685130 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.694885 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.702770 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.710724 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.718534 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.731279 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.739204 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:03Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.752508 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6ng6\" (UniqueName: \"kubernetes.io/projected/d360e6c4-2b40-4214-bb7c-5d08038c1b62-kube-api-access-k6ng6\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.752577 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.759448 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.759473 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.759480 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.759493 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.759501 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.814645 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.814702 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:03 crc kubenswrapper[4667]: E0929 17:10:03.814786 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.814801 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:03 crc kubenswrapper[4667]: E0929 17:10:03.814898 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:03 crc kubenswrapper[4667]: E0929 17:10:03.814951 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.853109 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.853166 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6ng6\" (UniqueName: \"kubernetes.io/projected/d360e6c4-2b40-4214-bb7c-5d08038c1b62-kube-api-access-k6ng6\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:03 crc kubenswrapper[4667]: E0929 17:10:03.853221 4667 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:03 crc kubenswrapper[4667]: E0929 17:10:03.853285 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs podName:d360e6c4-2b40-4214-bb7c-5d08038c1b62 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:04.353269179 +0000 UTC m=+32.851115948 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs") pod "network-metrics-daemon-cl5p9" (UID: "d360e6c4-2b40-4214-bb7c-5d08038c1b62") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.860719 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.860744 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.860752 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.860763 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.860773 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.865004 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6ng6\" (UniqueName: \"kubernetes.io/projected/d360e6c4-2b40-4214-bb7c-5d08038c1b62-kube-api-access-k6ng6\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.962427 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.962454 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.962463 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.962474 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.962483 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:03Z","lastTransitionTime":"2025-09-29T17:10:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.994614 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" event={"ID":"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a","Type":"ContainerStarted","Data":"e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e"} Sep 29 17:10:03 crc kubenswrapper[4667]: I0929 17:10:03.994653 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" event={"ID":"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a","Type":"ContainerStarted","Data":"c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.003291 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.010732 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.019770 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.027965 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.035534 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.042963 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.050632 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.056744 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.063959 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.064179 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.064209 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.064219 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.064229 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.064236 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.071309 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.078882 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.090010 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.096596 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.104436 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.110861 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.116783 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.166113 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.166139 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.166149 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.166164 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.166174 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.268026 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.268062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.268071 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.268083 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.268091 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.357870 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:04 crc kubenswrapper[4667]: E0929 17:10:04.357998 4667 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:04 crc kubenswrapper[4667]: E0929 17:10:04.358043 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs podName:d360e6c4-2b40-4214-bb7c-5d08038c1b62 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:05.358031863 +0000 UTC m=+33.855878632 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs") pod "network-metrics-daemon-cl5p9" (UID: "d360e6c4-2b40-4214-bb7c-5d08038c1b62") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.370111 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.370138 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.370148 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.370160 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.370169 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.471924 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.471950 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.471957 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.471968 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.471977 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.572950 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.572976 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.572986 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.572998 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.573006 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.674151 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.674178 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.674186 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.674196 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.674205 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.776565 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.776692 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.776700 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.776711 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.776719 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.878561 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.878584 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.878591 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.878600 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.878607 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.980316 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.980353 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.980361 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.980375 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:04 crc kubenswrapper[4667]: I0929 17:10:04.980384 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:04Z","lastTransitionTime":"2025-09-29T17:10:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.081797 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.081832 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.081855 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.081868 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.081877 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.183544 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.183587 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.183597 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.183611 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.183622 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.285150 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.285184 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.285194 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.285207 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.285216 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.367069 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.367177 4667 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.367240 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs podName:d360e6c4-2b40-4214-bb7c-5d08038c1b62 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:07.367224887 +0000 UTC m=+35.865071656 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs") pod "network-metrics-daemon-cl5p9" (UID: "d360e6c4-2b40-4214-bb7c-5d08038c1b62") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.386778 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.386809 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.386817 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.386829 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.386837 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.467422 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.467561 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:10:21.467548836 +0000 UTC m=+49.965395605 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.488379 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.488404 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.488413 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.488422 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.488429 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.568627 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.568658 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.568680 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.568696 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568727 4667 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568764 4667 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568774 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:21.568762386 +0000 UTC m=+50.066609165 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568793 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:21.56878622 +0000 UTC m=+50.066632989 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568803 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568823 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568834 4667 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568892 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568925 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568936 4667 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568900 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:21.568888432 +0000 UTC m=+50.066735201 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.568992 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:21.568979662 +0000 UTC m=+50.066826432 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.589766 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.589797 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.589805 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.589815 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.589822 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.691854 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.691886 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.691895 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.691907 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.691915 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.793272 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.793304 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.793313 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.793324 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.793333 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.815154 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.815189 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.815226 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.815168 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.815272 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.815357 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.815413 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:05 crc kubenswrapper[4667]: E0929 17:10:05.815451 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.894499 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.894536 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.894544 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.894555 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.894563 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.996325 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.996353 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.996363 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.996374 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:05 crc kubenswrapper[4667]: I0929 17:10:05.996382 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:05Z","lastTransitionTime":"2025-09-29T17:10:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.098391 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.098419 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.098427 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.098437 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.098444 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:06Z","lastTransitionTime":"2025-09-29T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.200195 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.200226 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.200235 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.200246 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.200255 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:06Z","lastTransitionTime":"2025-09-29T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.302360 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.302393 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.302401 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.302412 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.302420 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:06Z","lastTransitionTime":"2025-09-29T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.404405 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.404431 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.404439 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.404449 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.404458 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:06Z","lastTransitionTime":"2025-09-29T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.506053 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.506083 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.506092 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.506104 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.506113 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:06Z","lastTransitionTime":"2025-09-29T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.607610 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.607637 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.607644 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.607655 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.607663 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:06Z","lastTransitionTime":"2025-09-29T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.709023 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.709051 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.709061 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.709072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.709080 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:06Z","lastTransitionTime":"2025-09-29T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.810708 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.810726 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.810733 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.810742 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.810749 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:06Z","lastTransitionTime":"2025-09-29T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.912564 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.912594 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.912602 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.912613 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:06 crc kubenswrapper[4667]: I0929 17:10:06.912620 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:06Z","lastTransitionTime":"2025-09-29T17:10:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.014188 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.014223 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.014239 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.014257 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.014267 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.116092 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.116117 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.116125 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.116137 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.116147 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.217857 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.217883 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.217891 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.217900 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.217906 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.319556 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.319593 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.319602 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.319614 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.319622 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.387603 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:07 crc kubenswrapper[4667]: E0929 17:10:07.387716 4667 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:07 crc kubenswrapper[4667]: E0929 17:10:07.387766 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs podName:d360e6c4-2b40-4214-bb7c-5d08038c1b62 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:11.387751296 +0000 UTC m=+39.885598065 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs") pod "network-metrics-daemon-cl5p9" (UID: "d360e6c4-2b40-4214-bb7c-5d08038c1b62") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.421404 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.421427 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.421436 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.421446 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.421454 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.523337 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.523362 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.523369 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.523380 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.523388 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.625363 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.625385 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.625393 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.625403 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.625410 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.727731 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.727754 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.727762 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.727772 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.727780 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.815606 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:07 crc kubenswrapper[4667]: E0929 17:10:07.815680 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.815713 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.815792 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:07 crc kubenswrapper[4667]: E0929 17:10:07.815899 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.815968 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:07 crc kubenswrapper[4667]: E0929 17:10:07.816048 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:07 crc kubenswrapper[4667]: E0929 17:10:07.816162 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.829113 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.829148 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.829158 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.829170 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.829180 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.930540 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.930566 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.930574 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.930585 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:07 crc kubenswrapper[4667]: I0929 17:10:07.930594 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:07Z","lastTransitionTime":"2025-09-29T17:10:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.032570 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.032608 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.032619 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.032635 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.032652 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.134449 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.134478 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.134496 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.134507 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.134514 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.235723 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.235749 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.235758 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.235768 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.235775 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.337165 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.337201 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.337213 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.337225 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.337234 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.439064 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.439091 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.439101 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.439112 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.439122 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.525522 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.534572 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.540523 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.540542 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.540550 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.540560 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.540567 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.544237 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.553648 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.562706 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.571400 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.579573 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.587177 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.594290 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.600387 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.607393 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.615167 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.626272 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.633372 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.640995 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.641950 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.641984 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.641993 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.642005 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.642015 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.647879 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.656411 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:08Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.744253 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.744615 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.744674 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.744743 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.744806 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.846702 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.846734 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.846743 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.846754 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.846762 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.948168 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.948207 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.948218 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.948234 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:08 crc kubenswrapper[4667]: I0929 17:10:08.948245 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:08Z","lastTransitionTime":"2025-09-29T17:10:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.050164 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.050185 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.050193 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.050202 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.050209 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.152215 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.152243 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.152251 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.152261 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.152268 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.254277 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.254403 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.254486 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.254548 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.254608 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.356054 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.356255 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.356322 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.356381 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.356438 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.458170 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.458200 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.458210 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.458220 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.458227 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.559821 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.559934 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.559946 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.559958 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.559967 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.661702 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.661734 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.661744 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.661757 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.661765 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.763278 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.763299 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.763306 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.763315 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.763322 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.815940 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:09 crc kubenswrapper[4667]: E0929 17:10:09.816014 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.815942 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.816046 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:09 crc kubenswrapper[4667]: E0929 17:10:09.816120 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:09 crc kubenswrapper[4667]: E0929 17:10:09.816198 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.816217 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:09 crc kubenswrapper[4667]: E0929 17:10:09.816290 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.865141 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.865175 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.865186 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.865199 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.865209 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.967140 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.967192 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.967203 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.967215 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:09 crc kubenswrapper[4667]: I0929 17:10:09.967225 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:09Z","lastTransitionTime":"2025-09-29T17:10:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.069218 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.069248 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.069258 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.069269 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.069277 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.171276 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.171300 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.171308 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.171319 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.171329 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.273298 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.273331 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.273360 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.273374 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.273382 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.375345 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.375372 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.375381 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.375390 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.375398 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.476976 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.476998 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.477006 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.477014 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.477039 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.578630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.578656 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.578664 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.578675 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.578682 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.680181 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.680217 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.680226 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.680239 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.680249 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.782077 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.782108 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.782117 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.782129 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.782136 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.883646 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.883671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.883681 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.883692 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.883700 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.985859 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.985996 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.986079 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.986160 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:10 crc kubenswrapper[4667]: I0929 17:10:10.986223 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:10Z","lastTransitionTime":"2025-09-29T17:10:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.087835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.087894 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.087905 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.087917 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.087926 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:11Z","lastTransitionTime":"2025-09-29T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.189949 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.189974 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.189983 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.189993 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.190000 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:11Z","lastTransitionTime":"2025-09-29T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.291997 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.292033 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.292042 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.292054 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.292062 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:11Z","lastTransitionTime":"2025-09-29T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.394166 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.394197 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.394205 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.394218 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.394226 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:11Z","lastTransitionTime":"2025-09-29T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.419559 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:11 crc kubenswrapper[4667]: E0929 17:10:11.419675 4667 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:11 crc kubenswrapper[4667]: E0929 17:10:11.419713 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs podName:d360e6c4-2b40-4214-bb7c-5d08038c1b62 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:19.419702569 +0000 UTC m=+47.917549329 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs") pod "network-metrics-daemon-cl5p9" (UID: "d360e6c4-2b40-4214-bb7c-5d08038c1b62") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.496340 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.496378 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.496387 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.496400 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.496412 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:11Z","lastTransitionTime":"2025-09-29T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.598033 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.598066 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.598075 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.598086 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.598094 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:11Z","lastTransitionTime":"2025-09-29T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.699603 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.699635 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.699666 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.699679 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.699688 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:11Z","lastTransitionTime":"2025-09-29T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.801523 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.801553 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.801561 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.801572 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.801580 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:11Z","lastTransitionTime":"2025-09-29T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.815155 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.815197 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:11 crc kubenswrapper[4667]: E0929 17:10:11.815237 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.815250 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:11 crc kubenswrapper[4667]: E0929 17:10:11.815337 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.815376 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:11 crc kubenswrapper[4667]: E0929 17:10:11.815420 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:11 crc kubenswrapper[4667]: E0929 17:10:11.815493 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.824692 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.830898 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.837509 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.845145 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.852504 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.862060 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.870322 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.879035 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.886635 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.893918 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.900286 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.902596 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.902621 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.902630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.902641 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.902649 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:11Z","lastTransitionTime":"2025-09-29T17:10:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.907442 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.914278 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.921925 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.933404 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:11 crc kubenswrapper[4667]: I0929 17:10:11.940505 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.004756 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.004780 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.004788 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.004798 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.004806 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.106853 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.107049 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.107123 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.107210 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.107273 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.209306 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.209336 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.209344 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.209355 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.209362 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.311533 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.311682 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.311755 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.311836 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.311941 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.413978 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.414005 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.414013 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.414025 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.414032 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.515181 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.515216 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.515227 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.515241 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.515251 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.616718 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.616744 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.616755 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.616765 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.616773 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.718326 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.718351 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.718358 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.718367 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.718374 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.820162 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.820195 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.820204 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.820214 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.820224 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.921584 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.921614 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.921622 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.921632 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:12 crc kubenswrapper[4667]: I0929 17:10:12.921641 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:12Z","lastTransitionTime":"2025-09-29T17:10:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.023466 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.023503 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.023513 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.023525 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.023535 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.036496 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.036524 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.036532 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.036544 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.036551 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.044766 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.046980 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.047021 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.047031 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.047041 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.047048 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.054962 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.057480 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.057513 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.057525 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.057536 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.057545 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.065144 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.067181 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.067203 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.067211 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.067221 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.067228 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.074415 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.076372 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.076393 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.076401 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.076409 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.076416 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.083708 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:13Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.083806 4667 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.124918 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.125034 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.125104 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.125168 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.125229 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.226958 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.226990 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.227001 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.227012 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.227020 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.328780 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.328809 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.328819 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.328869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.328878 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.430593 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.430622 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.430631 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.430642 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.430649 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.532374 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.532402 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.532410 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.532421 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.532428 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.634237 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.634261 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.634269 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.634280 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.634288 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.736284 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.736314 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.736322 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.736333 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.736340 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.815364 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.815488 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.815600 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.815616 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.815778 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.815697 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.815647 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:13 crc kubenswrapper[4667]: E0929 17:10:13.816025 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.838141 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.838166 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.838174 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.838183 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.838190 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.939525 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.939552 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.939561 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.939571 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:13 crc kubenswrapper[4667]: I0929 17:10:13.939579 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:13Z","lastTransitionTime":"2025-09-29T17:10:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.041516 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.041642 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.041711 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.041783 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.041865 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.143794 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.144032 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.144108 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.144173 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.144235 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.245566 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.245595 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.245605 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.245618 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.245626 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.347551 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.347579 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.347587 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.347598 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.347607 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.449771 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.449798 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.449806 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.449818 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.449826 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.551172 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.551214 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.551222 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.551233 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.551241 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.652624 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.652646 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.652653 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.652663 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.652670 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.754814 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.754836 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.754860 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.754869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.754877 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.856682 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.856716 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.856727 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.856739 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.856748 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.958380 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.958406 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.958416 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.958436 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:14 crc kubenswrapper[4667]: I0929 17:10:14.958444 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:14Z","lastTransitionTime":"2025-09-29T17:10:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.060557 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.060671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.060748 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.060811 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.060891 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.162537 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.162711 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.162815 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.162933 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.163032 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.264651 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.264682 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.264693 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.264709 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.264718 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.366157 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.366184 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.366191 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.366200 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.366208 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.467729 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.467782 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.467791 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.467802 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.467811 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.569769 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.569910 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.569976 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.570032 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.570086 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.671869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.671895 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.671903 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.671915 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.671923 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.774005 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.774039 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.774050 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.774061 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.774068 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.815551 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:15 crc kubenswrapper[4667]: E0929 17:10:15.815642 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.815772 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.815868 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.815805 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:15 crc kubenswrapper[4667]: E0929 17:10:15.816066 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:15 crc kubenswrapper[4667]: E0929 17:10:15.816147 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:15 crc kubenswrapper[4667]: E0929 17:10:15.816194 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.876081 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.876117 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.876128 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.876138 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.876146 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.977919 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.978089 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.978156 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.978216 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:15 crc kubenswrapper[4667]: I0929 17:10:15.978281 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:15Z","lastTransitionTime":"2025-09-29T17:10:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.079447 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.079575 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.079645 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.079700 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.079758 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.182021 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.182057 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.182068 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.182079 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.182087 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.283967 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.284019 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.284028 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.284037 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.284044 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.385823 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.385900 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.385911 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.385924 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.385933 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.487284 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.487318 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.487329 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.487341 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.487349 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.589389 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.589428 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.589437 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.589447 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.589455 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.691465 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.691494 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.691501 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.691512 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.691524 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.793687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.793719 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.793727 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.793739 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.793748 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.894818 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.894868 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.894879 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.894891 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.894899 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.996718 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.996752 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.996760 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.996793 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:16 crc kubenswrapper[4667]: I0929 17:10:16.996803 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:16Z","lastTransitionTime":"2025-09-29T17:10:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.098898 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.098922 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.098930 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.098940 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.098948 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:17Z","lastTransitionTime":"2025-09-29T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.200678 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.200708 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.200717 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.200727 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.200734 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:17Z","lastTransitionTime":"2025-09-29T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.302975 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.303024 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.303035 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.303059 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.303068 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:17Z","lastTransitionTime":"2025-09-29T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.405040 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.405074 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.405083 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.405095 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.405103 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:17Z","lastTransitionTime":"2025-09-29T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.507536 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.507570 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.507578 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.507589 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.507597 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:17Z","lastTransitionTime":"2025-09-29T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.609141 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.609163 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.609171 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.609180 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.609188 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:17Z","lastTransitionTime":"2025-09-29T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.710639 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.710677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.710686 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.710700 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.710710 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:17Z","lastTransitionTime":"2025-09-29T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.812644 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.812669 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.812677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.812687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.812698 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:17Z","lastTransitionTime":"2025-09-29T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.815044 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.815074 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.815047 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.815180 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:17 crc kubenswrapper[4667]: E0929 17:10:17.815136 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:17 crc kubenswrapper[4667]: E0929 17:10:17.815262 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:17 crc kubenswrapper[4667]: E0929 17:10:17.815437 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:17 crc kubenswrapper[4667]: E0929 17:10:17.815731 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.815866 4667 scope.go:117] "RemoveContainer" containerID="b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.914141 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.914295 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.914304 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.914316 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:17 crc kubenswrapper[4667]: I0929 17:10:17.914324 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:17Z","lastTransitionTime":"2025-09-29T17:10:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.016047 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.016075 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.016084 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.016097 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.016105 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.025465 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/1.log" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.027218 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.027311 4667 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.041125 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.056512 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.067473 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.081414 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.092622 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.101895 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.109807 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.117413 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.118324 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.118357 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.118367 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.118379 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.118387 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.124116 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.131049 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.139898 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.148514 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.155790 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.163639 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.175470 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.182904 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:18Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.220298 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.220321 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.220330 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.220342 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.220350 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.322160 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.322191 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.322201 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.322213 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.322221 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.424389 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.424426 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.424434 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.424445 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.424454 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.525820 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.525859 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.525868 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.525879 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.525888 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.627697 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.627725 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.627735 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.627747 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.627756 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.729790 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.729824 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.729833 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.729871 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.729880 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.779202 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.831648 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.831678 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.831687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.831699 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.831713 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.933870 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.933898 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.933906 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.933915 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:18 crc kubenswrapper[4667]: I0929 17:10:18.933922 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:18Z","lastTransitionTime":"2025-09-29T17:10:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.030677 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/2.log" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.031106 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/1.log" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.033712 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b" exitCode=1 Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.033742 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.033766 4667 scope.go:117] "RemoveContainer" containerID="b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.034200 4667 scope.go:117] "RemoveContainer" containerID="4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b" Sep 29 17:10:19 crc kubenswrapper[4667]: E0929 17:10:19.034314 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.035088 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.035436 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.035504 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.035569 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.035623 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.044719 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.053098 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.062210 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.070634 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.078214 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.085504 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.092447 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.100104 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.112195 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b768d45cb18f43175697a375070a1bb4195ddebb29a18bb430594a15519a0d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:00Z\\\",\\\"message\\\":\\\"nt Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:00Z is after 2025-08-24T17:21:41Z]\\\\nI0929 17:10:00.576953 6086 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/package-server-manager-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"f9232b32-e89f-4c8e-acc4-c6801b70dcb0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/package-server-manager-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.L\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.119151 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.126541 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.132875 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.137134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.137220 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.137245 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.137259 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.137269 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.138725 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.146428 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.155522 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.164832 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:19Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.239167 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.239285 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.239350 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.239426 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.239479 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.341069 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.341099 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.341111 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.341123 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.341132 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.443210 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.443242 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.443250 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.443261 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.443269 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.489527 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:19 crc kubenswrapper[4667]: E0929 17:10:19.489628 4667 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:19 crc kubenswrapper[4667]: E0929 17:10:19.489665 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs podName:d360e6c4-2b40-4214-bb7c-5d08038c1b62 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:35.489653358 +0000 UTC m=+63.987500126 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs") pod "network-metrics-daemon-cl5p9" (UID: "d360e6c4-2b40-4214-bb7c-5d08038c1b62") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.544925 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.545028 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.545103 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.545166 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.545238 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.646603 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.646629 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.646637 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.646648 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.646657 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.748509 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.748533 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.748541 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.748549 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.748556 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.814887 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.814977 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.814899 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.814937 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:19 crc kubenswrapper[4667]: E0929 17:10:19.815080 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:19 crc kubenswrapper[4667]: E0929 17:10:19.815133 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:19 crc kubenswrapper[4667]: E0929 17:10:19.815172 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:19 crc kubenswrapper[4667]: E0929 17:10:19.815203 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.849902 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.849929 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.849938 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.849948 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.849959 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.951622 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.951648 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.951655 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.951664 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:19 crc kubenswrapper[4667]: I0929 17:10:19.951671 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:19Z","lastTransitionTime":"2025-09-29T17:10:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.036924 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/2.log" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.039248 4667 scope.go:117] "RemoveContainer" containerID="4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b" Sep 29 17:10:20 crc kubenswrapper[4667]: E0929 17:10:20.039366 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.048528 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.052519 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.052548 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.052557 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.052569 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.052577 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.056325 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.064343 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.070932 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.078115 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.086252 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.094190 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.100916 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.108366 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.120397 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.128532 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.135030 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.141445 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.149907 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.154480 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.154505 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.154513 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.154524 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.154532 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.157775 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.165535 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:20Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.256248 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.256273 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.256282 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.256294 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.256302 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.358329 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.358356 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.358364 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.358373 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.358391 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.460558 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.460591 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.460602 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.460613 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.460625 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.562828 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.562877 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.562886 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.562898 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.562907 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.664630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.664685 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.664694 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.664706 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.664714 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.766737 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.766770 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.766778 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.766792 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.766802 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.869064 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.869092 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.869100 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.869110 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.869117 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.970908 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.970939 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.970948 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.970961 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:20 crc kubenswrapper[4667]: I0929 17:10:20.970969 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:20Z","lastTransitionTime":"2025-09-29T17:10:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.072336 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.072365 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.072386 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.072396 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.072404 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.173683 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.173705 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.173712 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.173722 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.173729 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.274755 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.274782 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.274790 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.274799 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.274806 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.377178 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.377208 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.377216 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.377228 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.377236 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.478651 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.478685 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.478693 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.478705 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.478714 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.504025 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.504171 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:10:53.504154541 +0000 UTC m=+82.002001310 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.580204 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.580235 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.580243 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.580255 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.580264 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.604651 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.604685 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.604720 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.604738 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604816 4667 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604826 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604860 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604817 4667 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604871 4667 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604822 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604900 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604901 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:53.604888449 +0000 UTC m=+82.102735218 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604910 4667 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604918 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:53.604912124 +0000 UTC m=+82.102758893 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604933 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:53.604923887 +0000 UTC m=+82.102770646 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.604944 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 17:10:53.604938725 +0000 UTC m=+82.102785494 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.682063 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.682098 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.682106 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.682120 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.682129 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.784425 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.784465 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.784475 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.784487 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.784495 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.814851 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.814940 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.814950 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.814977 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.814983 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.815021 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.815092 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:21 crc kubenswrapper[4667]: E0929 17:10:21.815137 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.823755 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.832423 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.840573 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.848775 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.856556 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.863713 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.870997 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.878338 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.885523 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.886163 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.886192 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.886201 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.886213 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.886220 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.897141 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.904440 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.910812 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.918507 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.925989 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.934630 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.941803 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:21Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.988006 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.988534 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.988608 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.988674 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:21 crc kubenswrapper[4667]: I0929 17:10:21.988762 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:21Z","lastTransitionTime":"2025-09-29T17:10:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.090296 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.090328 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.090338 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.090351 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.090371 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:22Z","lastTransitionTime":"2025-09-29T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.191821 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.191864 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.191872 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.191885 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.191893 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:22Z","lastTransitionTime":"2025-09-29T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.293698 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.293832 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.293934 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.293998 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.294088 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:22Z","lastTransitionTime":"2025-09-29T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.395330 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.395354 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.395373 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.395382 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.395389 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:22Z","lastTransitionTime":"2025-09-29T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.496620 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.496654 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.496663 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.496675 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.496684 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:22Z","lastTransitionTime":"2025-09-29T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.598635 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.598732 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.598813 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.598909 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.598975 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:22Z","lastTransitionTime":"2025-09-29T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.700390 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.700424 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.700434 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.700448 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.700458 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:22Z","lastTransitionTime":"2025-09-29T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.802524 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.802550 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.802559 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.802569 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.802576 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:22Z","lastTransitionTime":"2025-09-29T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.904343 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.904445 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.904517 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.904575 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:22 crc kubenswrapper[4667]: I0929 17:10:22.904627 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:22Z","lastTransitionTime":"2025-09-29T17:10:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.005958 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.005984 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.005993 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.006003 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.006027 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.107686 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.107711 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.107719 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.107729 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.107735 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.209861 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.209973 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.210038 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.210099 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.210166 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.215780 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.215806 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.215814 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.215822 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.215829 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.223812 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:23Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.225813 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.225837 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.225858 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.225868 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.225874 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.232978 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:23Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.234780 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.234862 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.234887 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.234905 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.234916 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.242150 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:23Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.244069 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.244094 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.244102 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.244112 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.244121 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.251050 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:23Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.254250 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.254290 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.254301 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.254312 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.254322 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.262092 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:23Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.262201 4667 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.311494 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.311517 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.311527 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.311538 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.311546 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.413253 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.413280 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.413289 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.413298 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.413305 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.514909 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.514930 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.514938 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.514947 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.514955 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.616621 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.616650 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.616659 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.616671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.616679 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.718522 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.718560 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.718570 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.718584 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.718593 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.815618 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.815682 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.815713 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.815722 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.815771 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.815767 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.815828 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:23 crc kubenswrapper[4667]: E0929 17:10:23.815902 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.820109 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.820136 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.820146 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.820156 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.820170 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.921469 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.921497 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.921505 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.921515 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:23 crc kubenswrapper[4667]: I0929 17:10:23.921540 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:23Z","lastTransitionTime":"2025-09-29T17:10:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.023413 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.023437 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.023445 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.023454 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.023462 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.124652 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.124702 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.124710 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.124719 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.124726 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.226149 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.226192 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.226201 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.226210 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.226218 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.327456 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.327486 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.327493 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.327503 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.327510 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.429571 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.429618 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.429636 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.429646 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.429654 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.531079 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.531122 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.531131 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.531142 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.531150 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.633020 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.633053 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.633062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.633075 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.633085 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.720160 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.726586 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.727764 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.735197 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.735222 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.735230 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.735239 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.735263 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.736251 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.744484 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.753402 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.760488 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.768675 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.776648 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.784833 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.793573 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.799883 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.806772 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.814970 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.829141 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.836756 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.836753 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.836785 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.836912 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.836933 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.836942 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.845311 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.852109 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:24Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.938873 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.938910 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.938920 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.938933 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:24 crc kubenswrapper[4667]: I0929 17:10:24.938944 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:24Z","lastTransitionTime":"2025-09-29T17:10:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.040556 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.040594 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.040602 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.040613 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.040621 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.142921 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.142960 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.142970 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.142983 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.142992 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.245125 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.245153 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.245162 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.245175 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.245183 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.346861 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.346888 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.346897 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.346907 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.346914 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.448679 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.448704 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.448712 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.448723 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.448731 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.550062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.550089 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.550097 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.550107 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.550115 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.651877 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.651907 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.651915 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.651925 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.651932 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.753908 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.753931 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.753938 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.753948 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.753955 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.814604 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.814680 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:25 crc kubenswrapper[4667]: E0929 17:10:25.814690 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.814759 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:25 crc kubenswrapper[4667]: E0929 17:10:25.814836 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.814885 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:25 crc kubenswrapper[4667]: E0929 17:10:25.814920 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:25 crc kubenswrapper[4667]: E0929 17:10:25.814967 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.856036 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.856068 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.856078 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.856097 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.856106 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.957976 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.958020 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.958030 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.958043 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:25 crc kubenswrapper[4667]: I0929 17:10:25.958051 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:25Z","lastTransitionTime":"2025-09-29T17:10:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.059298 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.059337 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.059368 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.059378 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.059387 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.161561 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.161584 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.161592 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.161603 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.161611 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.263630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.263660 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.263669 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.263680 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.263688 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.365514 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.365542 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.365551 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.365562 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.365569 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.467193 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.467217 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.467225 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.467235 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.467242 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.568505 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.568526 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.568533 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.568544 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.568551 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.670002 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.670037 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.670047 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.670060 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.670071 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.771761 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.771790 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.771799 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.771811 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.771819 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.874881 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.875109 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.875172 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.875241 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.875310 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.976869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.976900 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.976909 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.976921 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:26 crc kubenswrapper[4667]: I0929 17:10:26.976931 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:26Z","lastTransitionTime":"2025-09-29T17:10:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.078965 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.079015 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.079030 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.079048 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.079064 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.180659 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.180691 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.180699 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.180710 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.180719 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.282087 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.282199 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.282210 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.282227 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.282234 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.384011 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.384039 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.384060 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.384072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.384080 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.485696 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.485723 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.485731 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.485740 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.485747 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.587722 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.587891 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.587951 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.588038 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.588102 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.689584 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.689614 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.689622 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.689632 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.689640 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.791486 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.791522 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.791534 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.791552 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.791561 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.814991 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:27 crc kubenswrapper[4667]: E0929 17:10:27.815089 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.815129 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.815156 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.815132 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:27 crc kubenswrapper[4667]: E0929 17:10:27.815209 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:27 crc kubenswrapper[4667]: E0929 17:10:27.815293 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:27 crc kubenswrapper[4667]: E0929 17:10:27.815498 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.893000 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.893116 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.893182 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.893261 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.893334 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.995234 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.995425 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.995491 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.995566 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:27 crc kubenswrapper[4667]: I0929 17:10:27.995626 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:27Z","lastTransitionTime":"2025-09-29T17:10:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.097278 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.097829 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.097945 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.098017 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.098079 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:28Z","lastTransitionTime":"2025-09-29T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.199776 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.199810 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.199819 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.199831 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.199839 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:28Z","lastTransitionTime":"2025-09-29T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.301076 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.301106 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.301115 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.301126 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.301135 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:28Z","lastTransitionTime":"2025-09-29T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.402921 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.402941 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.402949 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.402958 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.402964 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:28Z","lastTransitionTime":"2025-09-29T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.504552 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.504580 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.504590 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.504605 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.504615 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:28Z","lastTransitionTime":"2025-09-29T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.606121 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.606148 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.606157 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.606167 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.606176 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:28Z","lastTransitionTime":"2025-09-29T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.707883 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.707933 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.707944 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.707954 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.707963 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:28Z","lastTransitionTime":"2025-09-29T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.809960 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.810083 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.810168 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.810247 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.810341 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:28Z","lastTransitionTime":"2025-09-29T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.911615 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.911639 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.911647 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.911659 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:28 crc kubenswrapper[4667]: I0929 17:10:28.911666 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:28Z","lastTransitionTime":"2025-09-29T17:10:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.013308 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.013329 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.013337 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.013346 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.013353 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.114620 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.114647 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.114655 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.114665 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.114673 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.216288 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.216401 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.216459 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.216515 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.216586 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.318712 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.318747 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.318758 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.318771 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.318780 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.421042 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.421081 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.421089 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.421101 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.421109 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.522632 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.522664 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.522676 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.522692 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.522704 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.624518 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.624551 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.624561 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.624572 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.624580 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.726466 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.726510 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.726520 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.726532 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.726541 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.815391 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.815438 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.815465 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:29 crc kubenswrapper[4667]: E0929 17:10:29.815559 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.815580 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:29 crc kubenswrapper[4667]: E0929 17:10:29.815643 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:29 crc kubenswrapper[4667]: E0929 17:10:29.815729 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:29 crc kubenswrapper[4667]: E0929 17:10:29.815796 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.828134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.828169 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.828178 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.828187 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.828194 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.929191 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.929220 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.929228 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.929237 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:29 crc kubenswrapper[4667]: I0929 17:10:29.929244 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:29Z","lastTransitionTime":"2025-09-29T17:10:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.030695 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.030722 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.030768 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.030779 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.030786 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.132266 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.132301 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.132310 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.132320 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.132327 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.233464 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.233595 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.233671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.233737 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.233795 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.335545 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.335570 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.335579 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.335591 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.335598 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.437369 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.437392 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.437402 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.437412 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.437421 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.539130 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.539162 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.539171 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.539184 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.539194 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.640867 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.640896 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.640910 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.640921 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.640929 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.742776 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.742798 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.742806 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.742815 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.742822 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.844247 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.844270 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.844278 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.844296 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.844303 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.945431 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.945462 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.945471 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.945481 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:30 crc kubenswrapper[4667]: I0929 17:10:30.945489 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:30Z","lastTransitionTime":"2025-09-29T17:10:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.046863 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.046976 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.047044 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.047115 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.047181 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.149015 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.149048 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.149057 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.149068 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.149078 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.250251 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.250294 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.250305 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.250317 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.250326 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.351781 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.351811 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.351819 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.351829 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.351837 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.453454 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.453482 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.453490 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.453500 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.453507 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.554727 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.554754 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.554761 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.554770 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.554785 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.656068 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.656105 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.656116 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.656129 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.656139 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.757365 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.757395 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.757404 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.757413 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.757420 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.815596 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:31 crc kubenswrapper[4667]: E0929 17:10:31.815672 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.815713 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:31 crc kubenswrapper[4667]: E0929 17:10:31.815784 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.815713 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.815855 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:31 crc kubenswrapper[4667]: E0929 17:10:31.816255 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:31 crc kubenswrapper[4667]: E0929 17:10:31.816293 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.816470 4667 scope.go:117] "RemoveContainer" containerID="4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b" Sep 29 17:10:31 crc kubenswrapper[4667]: E0929 17:10:31.816616 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.825219 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.833638 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.841627 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.848758 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.854876 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.859168 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.859269 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.859355 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.859432 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.859498 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.862482 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.869471 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.876988 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.889155 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.896114 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.902956 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.909410 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.917109 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.923366 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.930602 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.939179 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.947168 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:31Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.961264 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.961299 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.961308 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.961321 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:31 crc kubenswrapper[4667]: I0929 17:10:31.961328 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:31Z","lastTransitionTime":"2025-09-29T17:10:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.062418 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.062450 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.062459 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.062471 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.062480 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.164062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.164095 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.164106 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.164119 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.164126 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.265776 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.265810 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.265818 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.265830 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.265838 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.367374 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.367400 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.367410 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.367419 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.367426 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.469673 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.469695 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.469704 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.469715 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.469722 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.572051 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.572116 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.572126 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.572137 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.572145 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.674346 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.674400 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.674412 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.674427 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.674435 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.776330 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.776372 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.776382 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.776397 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.776406 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.878797 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.878824 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.878832 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.878856 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.878864 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.980488 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.980522 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.980532 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.980545 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:32 crc kubenswrapper[4667]: I0929 17:10:32.980553 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:32Z","lastTransitionTime":"2025-09-29T17:10:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.081564 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.081599 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.081607 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.081620 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.081629 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.183608 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.183652 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.183663 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.183677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.183692 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.286076 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.286124 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.286134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.286150 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.286162 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.388019 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.388053 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.388062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.388075 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.388085 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.489977 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.490167 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.490257 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.490345 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.490407 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.495774 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.495806 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.495817 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.495829 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.495837 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.504662 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:33Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.506906 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.507002 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.507072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.507135 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.507197 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.514825 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:33Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.516983 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.517002 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.517014 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.517025 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.517032 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.524553 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:33Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.526694 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.526789 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.526872 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.526929 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.526989 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.535054 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:33Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.537294 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.537331 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.537340 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.537352 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.537362 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.545726 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:33Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.545831 4667 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.592085 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.592204 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.592306 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.592388 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.592452 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.693868 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.693897 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.693908 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.693921 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.693930 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.796129 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.796283 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.796352 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.796409 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.796464 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.814878 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.814879 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.814995 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.815058 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.814884 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.815118 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.815308 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:33 crc kubenswrapper[4667]: E0929 17:10:33.815390 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.898102 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.898125 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.898135 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.898145 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.898154 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.999574 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.999596 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.999604 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.999613 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:33 crc kubenswrapper[4667]: I0929 17:10:33.999620 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:33Z","lastTransitionTime":"2025-09-29T17:10:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.100666 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.100698 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.100708 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.100718 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.100724 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:34Z","lastTransitionTime":"2025-09-29T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.202373 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.202402 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.202413 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.202427 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.202436 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:34Z","lastTransitionTime":"2025-09-29T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.304409 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.304433 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.304443 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.304453 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.304460 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:34Z","lastTransitionTime":"2025-09-29T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.406038 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.406062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.406072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.406081 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.406088 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:34Z","lastTransitionTime":"2025-09-29T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.507857 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.507901 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.507912 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.507922 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.507929 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:34Z","lastTransitionTime":"2025-09-29T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.609335 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.609389 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.609401 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.609414 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.609424 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:34Z","lastTransitionTime":"2025-09-29T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.710497 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.710555 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.710566 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.710577 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.710586 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:34Z","lastTransitionTime":"2025-09-29T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.812241 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.812279 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.812287 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.812298 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.812305 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:34Z","lastTransitionTime":"2025-09-29T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.914564 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.914598 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.914607 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.914619 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:34 crc kubenswrapper[4667]: I0929 17:10:34.914627 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:34Z","lastTransitionTime":"2025-09-29T17:10:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.015921 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.015952 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.015960 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.015973 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.015981 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.117456 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.117591 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.117664 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.117724 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.117780 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.219262 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.219669 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.219730 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.219798 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.219879 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.321775 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.321794 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.321801 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.321809 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.321817 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.423579 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.423610 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.423619 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.423631 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.423640 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.516366 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:35 crc kubenswrapper[4667]: E0929 17:10:35.516494 4667 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:35 crc kubenswrapper[4667]: E0929 17:10:35.516539 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs podName:d360e6c4-2b40-4214-bb7c-5d08038c1b62 nodeName:}" failed. No retries permitted until 2025-09-29 17:11:07.516527241 +0000 UTC m=+96.014374010 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs") pod "network-metrics-daemon-cl5p9" (UID: "d360e6c4-2b40-4214-bb7c-5d08038c1b62") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.524659 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.524685 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.524695 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.524705 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.524713 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.626331 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.626364 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.626373 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.626386 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.626394 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.728368 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.728398 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.728407 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.728418 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.728426 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.814964 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.815038 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:35 crc kubenswrapper[4667]: E0929 17:10:35.815074 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:35 crc kubenswrapper[4667]: E0929 17:10:35.815127 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.815180 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:35 crc kubenswrapper[4667]: E0929 17:10:35.815219 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.815327 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:35 crc kubenswrapper[4667]: E0929 17:10:35.815392 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.829728 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.829754 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.829761 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.829771 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.829778 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.931285 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.931312 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.931323 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.931333 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:35 crc kubenswrapper[4667]: I0929 17:10:35.931341 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:35Z","lastTransitionTime":"2025-09-29T17:10:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.032488 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.032514 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.032522 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.032534 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.032543 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.134123 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.134152 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.134162 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.134173 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.134182 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.235612 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.235640 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.235649 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.235658 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.235664 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.337566 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.337605 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.337615 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.337627 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.337636 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.439016 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.439043 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.439052 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.439062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.439070 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.540396 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.540426 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.540438 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.540450 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.540459 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.642366 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.642397 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.642406 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.642419 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.642427 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.743895 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.743920 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.743929 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.743939 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.743946 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.844974 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.845005 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.845014 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.845024 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.845032 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.946134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.946159 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.946168 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.946178 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:36 crc kubenswrapper[4667]: I0929 17:10:36.946185 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:36Z","lastTransitionTime":"2025-09-29T17:10:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.047688 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.047708 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.047716 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.047724 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.047731 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.149644 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.149687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.149698 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.149714 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.149724 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.251104 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.251137 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.251145 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.251157 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.251165 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.355239 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.355481 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.355493 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.355505 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.355515 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.456918 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.456951 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.456960 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.456973 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.456982 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.558452 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.558481 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.558490 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.558502 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.558511 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.659974 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.660001 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.660010 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.660021 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.660028 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.761173 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.761201 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.761209 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.761219 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.761238 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.814769 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.814814 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.814815 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.814777 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:37 crc kubenswrapper[4667]: E0929 17:10:37.814888 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:37 crc kubenswrapper[4667]: E0929 17:10:37.814961 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:37 crc kubenswrapper[4667]: E0929 17:10:37.815025 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:37 crc kubenswrapper[4667]: E0929 17:10:37.815079 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.862933 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.862964 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.862973 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.862983 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.862991 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.964943 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.964978 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.964986 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.964995 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:37 crc kubenswrapper[4667]: I0929 17:10:37.965003 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:37Z","lastTransitionTime":"2025-09-29T17:10:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.067009 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.067049 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.067059 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.067072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.067082 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.075786 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/0.log" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.075821 4667 generic.go:334] "Generic (PLEG): container finished" podID="788f770a-3181-4b66-981c-90ffb7fc49c0" containerID="1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c" exitCode=1 Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.075858 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2bf6m" event={"ID":"788f770a-3181-4b66-981c-90ffb7fc49c0","Type":"ContainerDied","Data":"1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.076106 4667 scope.go:117] "RemoveContainer" containerID="1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.084634 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.092346 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.101871 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.109276 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.117812 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.127315 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.135353 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.142951 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.149482 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.157604 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.165124 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.169189 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.169213 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.169230 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.169241 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.169250 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.173796 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:37Z\\\",\\\"message\\\":\\\"2025-09-29T17:09:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96\\\\n2025-09-29T17:09:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96 to /host/opt/cni/bin/\\\\n2025-09-29T17:09:52Z [verbose] multus-daemon started\\\\n2025-09-29T17:09:52Z [verbose] Readiness Indicator file check\\\\n2025-09-29T17:10:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.185247 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.192827 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.199662 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.206352 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.214044 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:38Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.270941 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.270971 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.270982 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.270996 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.271007 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.372319 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.372355 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.372365 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.372380 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.372389 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.474027 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.474053 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.474061 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.474069 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.474077 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.575787 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.575829 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.575837 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.575868 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.575878 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.677274 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.677308 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.677318 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.677332 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.677340 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.778887 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.778916 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.778927 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.778938 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.778945 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.881744 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.881789 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.881801 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.881817 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.881834 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.984133 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.984159 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.984167 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.984177 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:38 crc kubenswrapper[4667]: I0929 17:10:38.984186 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:38Z","lastTransitionTime":"2025-09-29T17:10:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.079736 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/0.log" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.079779 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2bf6m" event={"ID":"788f770a-3181-4b66-981c-90ffb7fc49c0","Type":"ContainerStarted","Data":"150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.085707 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.085736 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.085746 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.085759 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.085768 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:39Z","lastTransitionTime":"2025-09-29T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.089328 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.097814 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.106796 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.116382 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.124388 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.131929 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.138432 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.145950 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.153913 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.162926 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.170830 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.177556 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.185545 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:37Z\\\",\\\"message\\\":\\\"2025-09-29T17:09:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96\\\\n2025-09-29T17:09:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96 to /host/opt/cni/bin/\\\\n2025-09-29T17:09:52Z [verbose] multus-daemon started\\\\n2025-09-29T17:09:52Z [verbose] Readiness Indicator file check\\\\n2025-09-29T17:10:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.187444 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.187474 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.187485 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.187495 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.187503 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:39Z","lastTransitionTime":"2025-09-29T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.196594 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.203620 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.211278 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.217786 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:39Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.289597 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.289626 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.289652 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.289665 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.289673 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:39Z","lastTransitionTime":"2025-09-29T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.391741 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.391770 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.391784 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.391796 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.391805 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:39Z","lastTransitionTime":"2025-09-29T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.493463 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.493487 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.493495 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.493505 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.493530 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:39Z","lastTransitionTime":"2025-09-29T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.595389 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.595455 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.595467 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.595476 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.595483 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:39Z","lastTransitionTime":"2025-09-29T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.697134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.697155 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.697301 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.697315 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.697322 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:39Z","lastTransitionTime":"2025-09-29T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.799220 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.799256 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.799267 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.799281 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.799290 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:39Z","lastTransitionTime":"2025-09-29T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.815677 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.815695 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:39 crc kubenswrapper[4667]: E0929 17:10:39.815786 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:39 crc kubenswrapper[4667]: E0929 17:10:39.815873 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.815883 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:39 crc kubenswrapper[4667]: E0929 17:10:39.815962 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.816015 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:39 crc kubenswrapper[4667]: E0929 17:10:39.816159 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.901216 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.901275 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.901291 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.901311 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:39 crc kubenswrapper[4667]: I0929 17:10:39.901325 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:39Z","lastTransitionTime":"2025-09-29T17:10:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.003397 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.003422 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.003431 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.003440 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.003449 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.105622 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.105657 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.105666 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.105676 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.105683 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.207875 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.207903 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.207914 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.207924 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.207930 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.309990 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.310031 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.310045 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.310062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.310073 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.411765 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.411907 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.412000 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.412075 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.412138 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.513713 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.513745 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.513754 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.513766 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.513776 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.614960 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.614984 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.614993 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.615003 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.615027 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.716805 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.716937 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.717016 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.717079 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.717138 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.819118 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.819251 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.819324 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.819395 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.819460 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.920869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.920904 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.920913 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.920925 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:40 crc kubenswrapper[4667]: I0929 17:10:40.920934 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:40Z","lastTransitionTime":"2025-09-29T17:10:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.022512 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.022539 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.022548 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.022556 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.022564 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.123836 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.124002 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.124065 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.124140 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.124219 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.225669 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.225696 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.225705 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.225717 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.225725 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.327493 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.327526 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.327536 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.327547 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.327557 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.429049 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.429066 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.429076 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.429086 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.429093 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.530880 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.530900 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.530907 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.530917 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.530924 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.631951 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.631986 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.631996 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.632008 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.632032 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.733591 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.733615 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.733623 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.733633 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.733639 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.815324 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.815363 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.815375 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:41 crc kubenswrapper[4667]: E0929 17:10:41.815427 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:41 crc kubenswrapper[4667]: E0929 17:10:41.815473 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:41 crc kubenswrapper[4667]: E0929 17:10:41.815512 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.815631 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:41 crc kubenswrapper[4667]: E0929 17:10:41.815769 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.824133 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.830648 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.835505 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.835602 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.835677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.835746 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.835810 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.837709 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.845797 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.853534 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.862374 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.869409 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.879011 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.888436 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.896052 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.903892 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.910008 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.917169 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.923373 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.932213 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:37Z\\\",\\\"message\\\":\\\"2025-09-29T17:09:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96\\\\n2025-09-29T17:09:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96 to /host/opt/cni/bin/\\\\n2025-09-29T17:09:52Z [verbose] multus-daemon started\\\\n2025-09-29T17:09:52Z [verbose] Readiness Indicator file check\\\\n2025-09-29T17:10:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.938177 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.938209 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.938219 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.938233 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.938241 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:41Z","lastTransitionTime":"2025-09-29T17:10:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.943715 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:41 crc kubenswrapper[4667]: I0929 17:10:41.950594 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:41Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.039697 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.039731 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.039791 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.039825 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.039834 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.141424 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.141558 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.141629 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.141694 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.141766 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.242990 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.243016 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.243024 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.243035 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.243043 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.344893 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.344923 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.344932 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.344944 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.344952 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.446467 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.446495 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.446504 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.446517 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.446525 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.547998 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.548028 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.548037 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.548049 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.548058 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.649932 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.649967 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.649976 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.649989 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.649998 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.751941 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.751977 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.751987 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.752000 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.752009 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.815856 4667 scope.go:117] "RemoveContainer" containerID="4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.853320 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.853346 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.853355 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.853365 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.853375 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.955656 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.955691 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.955700 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.955712 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:42 crc kubenswrapper[4667]: I0929 17:10:42.955721 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:42Z","lastTransitionTime":"2025-09-29T17:10:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.056799 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.056827 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.056835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.056862 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.056872 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.089323 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/2.log" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.091528 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.091885 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.103321 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.111743 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.122803 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.130074 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.138626 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.148484 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.157580 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.158570 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.158593 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.158601 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.158612 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.158620 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.166639 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.174714 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.181446 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.188947 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.197112 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:37Z\\\",\\\"message\\\":\\\"2025-09-29T17:09:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96\\\\n2025-09-29T17:09:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96 to /host/opt/cni/bin/\\\\n2025-09-29T17:09:52Z [verbose] multus-daemon started\\\\n2025-09-29T17:09:52Z [verbose] Readiness Indicator file check\\\\n2025-09-29T17:10:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.214998 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.222275 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.231419 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.238646 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.246032 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.259912 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.259944 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.259952 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.259965 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.259973 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.361630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.361658 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.361666 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.361677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.361684 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.463690 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.463717 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.463726 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.463738 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.463747 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.565817 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.565869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.565881 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.565897 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.565907 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.667773 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.667795 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.667803 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.667813 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.667838 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.769531 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.769559 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.769568 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.769580 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.769587 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.815432 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.815465 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.815515 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.815438 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.815683 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.815781 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.815887 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.815960 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.864570 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.864603 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.864613 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.864624 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.864633 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.875204 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.877263 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.877305 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.877316 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.877328 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.877335 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.886241 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.888218 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.888245 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.888271 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.888282 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.888290 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.895604 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.897654 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.897682 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.897691 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.897718 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.897728 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.904896 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.906775 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.906802 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.906811 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.906820 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.906833 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.914681 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:43Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:43 crc kubenswrapper[4667]: E0929 17:10:43.914785 4667 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.915650 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.915675 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.915683 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.915692 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:43 crc kubenswrapper[4667]: I0929 17:10:43.915699 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:43Z","lastTransitionTime":"2025-09-29T17:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.017165 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.017194 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.017220 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.017239 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.017248 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.094148 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/3.log" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.094659 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/2.log" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.096311 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415" exitCode=1 Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.096337 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.096367 4667 scope.go:117] "RemoveContainer" containerID="4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.096833 4667 scope.go:117] "RemoveContainer" containerID="d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415" Sep 29 17:10:44 crc kubenswrapper[4667]: E0929 17:10:44.096956 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.105147 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.113404 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.118962 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.118986 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.118994 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.119022 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.119030 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.124636 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.140089 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.163256 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.174693 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.183731 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.192305 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.198817 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.206399 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.214011 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.220631 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.220660 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.220669 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.220681 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.220690 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.224811 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:37Z\\\",\\\"message\\\":\\\"2025-09-29T17:09:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96\\\\n2025-09-29T17:09:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96 to /host/opt/cni/bin/\\\\n2025-09-29T17:09:52Z [verbose] multus-daemon started\\\\n2025-09-29T17:09:52Z [verbose] Readiness Indicator file check\\\\n2025-09-29T17:10:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.237174 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4efa9c8f03478f928580aadae837daff33cc58e00ae6814233f3b1c25c21d56b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:18Z\\\",\\\"message\\\":\\\".lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.119\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI0929 17:10:18.405499 6337 services_controller.go:444] Built service openshift-multus/multus-admission-controller LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI0929 17:10:18.405285 6337 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}\\\\nF0929 17:10:18.405524 6337 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controll\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"365] Adding new object: *v1.Pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422928 6733 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-h7cv4 in node crc\\\\nI0929 17:10:43.422932 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-h7cv4 after 0 failed attempt(s)\\\\nI0929 17:10:43.422935 6733 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422809 6733 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nI0929 17:10:43.422941 6733 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-l8rmj in node crc\\\\nI0929 17:10:43.422945 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj after 0 failed attempt(s)\\\\nI0929 17:10:43.422949 6733 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nF0929 17:10:43.422795 6733 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller init\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.244615 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.253518 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.259754 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.266006 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:44Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.322899 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.322922 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.322930 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.322941 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.322949 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.424599 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.424642 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.424652 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.424662 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.424669 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.526024 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.526055 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.526065 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.526096 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.526106 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.627263 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.627293 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.627320 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.627332 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.627343 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.728718 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.728746 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.728754 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.728767 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.728775 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.830058 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.830088 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.830098 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.830109 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.830117 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.931807 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.931879 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.931887 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.931898 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:44 crc kubenswrapper[4667]: I0929 17:10:44.931906 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:44Z","lastTransitionTime":"2025-09-29T17:10:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.033801 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.033826 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.033835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.033861 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.033868 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.099209 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/3.log" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.101901 4667 scope.go:117] "RemoveContainer" containerID="d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415" Sep 29 17:10:45 crc kubenswrapper[4667]: E0929 17:10:45.102030 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.111497 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.120271 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.128080 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.135582 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.135624 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.135634 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.135644 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.135652 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.136089 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.144170 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.151527 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.157597 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.164638 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.171603 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.180702 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.188579 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.195664 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.203820 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:37Z\\\",\\\"message\\\":\\\"2025-09-29T17:09:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96\\\\n2025-09-29T17:09:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96 to /host/opt/cni/bin/\\\\n2025-09-29T17:09:52Z [verbose] multus-daemon started\\\\n2025-09-29T17:09:52Z [verbose] Readiness Indicator file check\\\\n2025-09-29T17:10:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.215791 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"365] Adding new object: *v1.Pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422928 6733 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-h7cv4 in node crc\\\\nI0929 17:10:43.422932 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-h7cv4 after 0 failed attempt(s)\\\\nI0929 17:10:43.422935 6733 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422809 6733 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nI0929 17:10:43.422941 6733 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-l8rmj in node crc\\\\nI0929 17:10:43.422945 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj after 0 failed attempt(s)\\\\nI0929 17:10:43.422949 6733 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nF0929 17:10:43.422795 6733 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller init\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.223563 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.230008 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.235897 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:45Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.236862 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.236888 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.236899 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.236911 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.236920 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.338806 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.338836 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.338860 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.338874 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.338882 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.440907 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.440936 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.440945 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.440958 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.440968 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.542537 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.542567 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.542578 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.542594 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.542602 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.644987 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.645016 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.645026 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.645037 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.645045 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.747091 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.747122 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.747131 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.747151 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.747160 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.815010 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:45 crc kubenswrapper[4667]: E0929 17:10:45.815087 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.815107 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.815160 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:45 crc kubenswrapper[4667]: E0929 17:10:45.815219 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.815292 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:45 crc kubenswrapper[4667]: E0929 17:10:45.815413 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:45 crc kubenswrapper[4667]: E0929 17:10:45.815304 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.848829 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.848872 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.848882 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.848893 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.848901 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.950078 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.950103 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.950111 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.950122 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:45 crc kubenswrapper[4667]: I0929 17:10:45.950140 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:45Z","lastTransitionTime":"2025-09-29T17:10:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.051752 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.051776 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.051784 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.051794 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.051800 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.153179 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.153206 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.153215 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.153224 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.153232 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.255113 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.255164 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.255174 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.255184 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.255192 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.357034 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.357060 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.357068 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.357078 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.357084 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.458807 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.458833 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.458857 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.458868 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.458876 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.561097 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.561118 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.561137 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.561149 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.561157 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.663137 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.663187 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.663196 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.663208 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.663219 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.765109 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.765155 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.765166 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.765179 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.765188 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.866652 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.866679 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.866687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.866698 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.866705 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.968791 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.968821 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.968832 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.968867 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:46 crc kubenswrapper[4667]: I0929 17:10:46.968876 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:46Z","lastTransitionTime":"2025-09-29T17:10:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.070717 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.070757 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.070766 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.070778 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.070786 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.172199 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.172226 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.172234 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.172243 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.172250 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.273532 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.273569 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.273577 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.273590 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.273599 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.375526 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.375549 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.375557 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.375566 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.375575 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.477595 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.477630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.477638 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.477647 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.477654 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.579273 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.579299 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.579307 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.579317 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.579324 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.681019 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.681048 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.681056 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.681069 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.681083 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.782355 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.782383 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.782392 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.782402 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.782410 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.814876 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.814897 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.814934 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:47 crc kubenswrapper[4667]: E0929 17:10:47.814959 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:47 crc kubenswrapper[4667]: E0929 17:10:47.815045 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.815059 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:47 crc kubenswrapper[4667]: E0929 17:10:47.815082 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:47 crc kubenswrapper[4667]: E0929 17:10:47.815130 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.884560 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.884585 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.884593 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.884603 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.884609 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.985976 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.986029 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.986040 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.986055 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:47 crc kubenswrapper[4667]: I0929 17:10:47.986065 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:47Z","lastTransitionTime":"2025-09-29T17:10:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.087359 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.087388 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.087397 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.087407 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.087415 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:48Z","lastTransitionTime":"2025-09-29T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.189503 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.189532 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.189541 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.189551 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.189563 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:48Z","lastTransitionTime":"2025-09-29T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.291778 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.291803 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.291812 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.291823 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.291830 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:48Z","lastTransitionTime":"2025-09-29T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.393546 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.393569 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.393577 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.393586 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.393594 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:48Z","lastTransitionTime":"2025-09-29T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.495251 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.495285 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.495293 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.495305 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.495313 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:48Z","lastTransitionTime":"2025-09-29T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.596762 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.596790 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.596799 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.596810 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.596817 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:48Z","lastTransitionTime":"2025-09-29T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.698605 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.698635 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.698646 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.698659 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.698668 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:48Z","lastTransitionTime":"2025-09-29T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.800683 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.800717 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.800726 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.800740 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.800749 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:48Z","lastTransitionTime":"2025-09-29T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.824508 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.902657 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.902687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.902694 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.902705 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:48 crc kubenswrapper[4667]: I0929 17:10:48.902714 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:48Z","lastTransitionTime":"2025-09-29T17:10:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.004328 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.004357 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.004366 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.004377 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.004384 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.105601 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.105630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.105638 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.105651 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.105661 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.206750 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.206785 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.206796 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.206808 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.206818 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.308923 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.308957 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.308967 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.308980 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.308989 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.410731 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.410763 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.410772 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.410783 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.410794 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.512386 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.512429 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.512438 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.512448 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.512457 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.613914 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.614052 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.614134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.614200 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.614251 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.716225 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.716251 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.716260 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.716271 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.716278 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.815493 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.815508 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.815547 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:49 crc kubenswrapper[4667]: E0929 17:10:49.815620 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.815636 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:49 crc kubenswrapper[4667]: E0929 17:10:49.815759 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:49 crc kubenswrapper[4667]: E0929 17:10:49.815796 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:49 crc kubenswrapper[4667]: E0929 17:10:49.815860 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.817970 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.817995 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.818004 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.818016 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.818024 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.919571 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.919605 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.919614 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.919625 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:49 crc kubenswrapper[4667]: I0929 17:10:49.919633 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:49Z","lastTransitionTime":"2025-09-29T17:10:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.021047 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.021099 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.021109 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.021120 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.021134 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.123188 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.123216 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.123225 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.123236 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.123245 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.225485 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.225510 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.225519 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.225542 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.225551 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.327267 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.327300 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.327309 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.327320 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.327330 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.429091 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.429122 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.429130 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.429142 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.429150 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.530740 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.530780 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.530788 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.530802 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.530812 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.632665 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.632695 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.632706 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.632723 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.632732 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.734406 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.734432 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.734441 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.734452 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.734460 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.835982 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.836017 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.836027 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.836039 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.836047 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.937580 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.937608 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.937616 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.937627 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:50 crc kubenswrapper[4667]: I0929 17:10:50.937635 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:50Z","lastTransitionTime":"2025-09-29T17:10:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.038661 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.038687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.038699 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.038710 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.038718 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.140299 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.140327 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.140337 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.140348 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.140356 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.242159 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.242194 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.242204 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.242216 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.242224 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.343878 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.343907 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.343915 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.343925 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.343935 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.445044 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.445075 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.445084 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.445094 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.445102 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.546909 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.546929 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.546938 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.546948 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.546957 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.649103 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.649132 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.649143 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.649157 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.649167 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.750710 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.750745 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.750755 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.750769 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.750779 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.815580 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.815678 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:51 crc kubenswrapper[4667]: E0929 17:10:51.815773 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.815894 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:51 crc kubenswrapper[4667]: E0929 17:10:51.815947 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:51 crc kubenswrapper[4667]: E0929 17:10:51.815996 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.815898 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:51 crc kubenswrapper[4667]: E0929 17:10:51.816104 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.825321 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.832710 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.839123 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.847281 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.852871 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.852895 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.852904 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.852917 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.852926 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.859871 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fc84a9a-2ed1-4304-b7e8-37d43fee7d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb7c287f124ce6b9696a6ae200279ba2e4509201c42176c5f1c21d78f948ea3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbfa2510696eb81866c2158c7eb7102a0c591f568ed85b48ba594d769ddd98d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d89da318419cde5f78e062ae14168f716c169fb0bb7a2a7468974021ff75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57073aea9b43bae01cf9e3f0e23390e2550bc3942f07cf94bdb9b92f12d4a9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6f02c341c7866bad9a9a485185b47b00d93f8d0b2f289d27f1ef851edbc18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fe7aa4a24221e184afdd808a3a04146b5fb2ef77c3720ca69a1358aff749618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe7aa4a24221e184afdd808a3a04146b5fb2ef77c3720ca69a1358aff749618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ed387e17159598d2c11525b5534226b01ac0f7beb08e43aed9d93024de185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b98ed387e17159598d2c11525b5534226b01ac0f7beb08e43aed9d93024de185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6ae461746f91f42609f4b85cf93ffcfd68aa9b865190a44af69532e600a55f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae461746f91f42609f4b85cf93ffcfd68aa9b865190a44af69532e600a55f46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.867312 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.877200 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.885272 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.894224 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.902263 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.911897 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.920136 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.926555 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.933458 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.941653 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.949437 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:37Z\\\",\\\"message\\\":\\\"2025-09-29T17:09:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96\\\\n2025-09-29T17:09:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96 to /host/opt/cni/bin/\\\\n2025-09-29T17:09:52Z [verbose] multus-daemon started\\\\n2025-09-29T17:09:52Z [verbose] Readiness Indicator file check\\\\n2025-09-29T17:10:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.954968 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.954996 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.955005 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.955017 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.955024 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:51Z","lastTransitionTime":"2025-09-29T17:10:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.960458 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"365] Adding new object: *v1.Pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422928 6733 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-h7cv4 in node crc\\\\nI0929 17:10:43.422932 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-h7cv4 after 0 failed attempt(s)\\\\nI0929 17:10:43.422935 6733 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422809 6733 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nI0929 17:10:43.422941 6733 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-l8rmj in node crc\\\\nI0929 17:10:43.422945 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj after 0 failed attempt(s)\\\\nI0929 17:10:43.422949 6733 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nF0929 17:10:43.422795 6733 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller init\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:51 crc kubenswrapper[4667]: I0929 17:10:51.967523 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:51Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.057026 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.057072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.057084 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.057110 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.057120 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.158610 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.158641 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.158651 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.158662 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.158670 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.260571 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.260605 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.260613 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.260625 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.260633 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.362802 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.362835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.362859 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.362871 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.362880 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.464869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.464896 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.464904 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.464915 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.464923 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.566486 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.566528 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.566536 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.566549 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.566558 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.668523 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.668550 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.668557 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.668568 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.668577 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.770810 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.770853 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.770864 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.770876 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.770884 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.872147 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.872174 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.872183 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.872192 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.872200 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.973709 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.973732 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.973739 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.973750 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:52 crc kubenswrapper[4667]: I0929 17:10:52.973758 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:52Z","lastTransitionTime":"2025-09-29T17:10:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.075698 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.075721 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.075729 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.075740 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.075747 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.177198 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.177233 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.177250 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.177263 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.177272 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.278412 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.278437 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.278445 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.278472 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.278486 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.380244 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.380331 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.380408 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.380463 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.380542 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.481524 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.481618 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.481676 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.481727 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.481778 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.553009 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.553097 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:57.553081023 +0000 UTC m=+146.050927803 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.583258 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.583593 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.583653 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.583711 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.583766 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.653753 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.653890 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.653969 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.654114 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.653890 4667 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654270 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:11:57.654259277 +0000 UTC m=+146.152106046 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.653921 4667 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654402 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-09-29 17:11:57.654395793 +0000 UTC m=+146.152242562 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654111 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654519 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654571 4667 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654639 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-09-29 17:11:57.654632757 +0000 UTC m=+146.152479527 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654149 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654734 4667 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654747 4667 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.654795 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-09-29 17:11:57.654785213 +0000 UTC m=+146.152631982 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.685204 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.685229 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.685236 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.685246 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.685253 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.786631 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.786662 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.786671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.786685 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.786694 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.814902 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.814921 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.814975 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.814904 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.815097 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.815140 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.815249 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.815285 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.888083 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.888224 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.888301 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.888365 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.888441 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.961087 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.961244 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.961303 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.961355 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.961403 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.970016 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.972456 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.972548 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.972602 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.972652 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.972711 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.980383 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.983478 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.983508 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.983518 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.983529 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.983540 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:53 crc kubenswrapper[4667]: E0929 17:10:53.992155 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:53Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.994290 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.994310 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.994319 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.994329 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:53 crc kubenswrapper[4667]: I0929 17:10:53.994338 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:53Z","lastTransitionTime":"2025-09-29T17:10:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: E0929 17:10:54.002007 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.004296 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.004322 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.004331 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.004342 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.004348 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: E0929 17:10:54.011589 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:10:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:10:54Z is after 2025-08-24T17:21:41Z" Sep 29 17:10:54 crc kubenswrapper[4667]: E0929 17:10:54.011686 4667 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.012537 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.012556 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.012564 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.012572 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.012579 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.114022 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.114060 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.114069 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.114080 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.114089 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.216126 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.216271 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.216352 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.216417 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.216474 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.317869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.317904 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.317912 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.317925 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.317933 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.419796 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.419827 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.419835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.419863 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.419871 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.521281 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.521314 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.521322 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.521332 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.521340 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.622898 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.622950 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.622959 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.622969 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.622977 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.724798 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.724825 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.724835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.724861 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.724870 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.826233 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.826252 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.826260 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.826270 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.826278 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.927720 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.927747 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.927755 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.927765 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:54 crc kubenswrapper[4667]: I0929 17:10:54.927773 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:54Z","lastTransitionTime":"2025-09-29T17:10:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.029408 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.029430 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.029438 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.029449 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.029456 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.130971 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.131006 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.131016 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.131042 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.131052 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.232953 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.232979 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.232986 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.232996 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.233003 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.334641 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.334658 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.334667 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.334676 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.334683 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.436326 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.436496 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.436558 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.436625 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.436684 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.538055 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.538165 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.538236 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.538293 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.538347 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.639955 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.639987 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.639996 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.640007 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.640014 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.741929 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.742152 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.742226 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.742295 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.742353 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.814989 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.815099 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.815099 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.815123 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:55 crc kubenswrapper[4667]: E0929 17:10:55.815434 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:55 crc kubenswrapper[4667]: E0929 17:10:55.815507 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:55 crc kubenswrapper[4667]: E0929 17:10:55.815584 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.815682 4667 scope.go:117] "RemoveContainer" containerID="d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415" Sep 29 17:10:55 crc kubenswrapper[4667]: E0929 17:10:55.815782 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:55 crc kubenswrapper[4667]: E0929 17:10:55.815823 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.844416 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.844439 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.844448 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.844462 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.844470 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.946604 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.946633 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.946642 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.946652 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:55 crc kubenswrapper[4667]: I0929 17:10:55.946660 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:55Z","lastTransitionTime":"2025-09-29T17:10:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.048569 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.048598 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.048607 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.048617 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.048623 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.150399 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.150422 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.150430 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.150438 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.150446 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.251787 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.252035 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.252104 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.252167 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.252223 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.354446 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.354610 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.354678 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.354744 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.354814 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.457436 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.457469 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.457477 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.457486 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.457494 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.559046 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.559072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.559080 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.559091 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.559098 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.660355 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.660376 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.660384 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.660393 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.660399 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.761755 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.761789 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.761797 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.761808 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.761816 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.863726 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.863747 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.863754 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.863764 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.863770 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.965393 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.965413 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.965422 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.965430 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:56 crc kubenswrapper[4667]: I0929 17:10:56.965436 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:56Z","lastTransitionTime":"2025-09-29T17:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.067013 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.067043 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.067051 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.067062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.067069 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.168766 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.168794 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.168802 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.168829 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.168852 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.270791 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.270826 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.270835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.270865 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.270873 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.372703 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.372729 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.372737 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.372764 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.372772 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.474663 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.474697 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.474706 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.474718 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.474726 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.575967 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.576004 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.576012 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.576022 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.576030 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.677780 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.677806 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.677813 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.677824 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.677832 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.779326 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.779353 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.779361 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.779370 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.779377 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.815056 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.815084 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.815141 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:57 crc kubenswrapper[4667]: E0929 17:10:57.815233 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.815264 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:57 crc kubenswrapper[4667]: E0929 17:10:57.815325 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:57 crc kubenswrapper[4667]: E0929 17:10:57.815408 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:57 crc kubenswrapper[4667]: E0929 17:10:57.815455 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.880892 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.880930 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.880937 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.880946 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.880953 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.982815 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.982839 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.982864 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.982874 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:57 crc kubenswrapper[4667]: I0929 17:10:57.982881 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:57Z","lastTransitionTime":"2025-09-29T17:10:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.084505 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.084528 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.084536 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.084544 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.084552 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:58Z","lastTransitionTime":"2025-09-29T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.186471 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.186492 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.186501 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.186512 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.186523 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:58Z","lastTransitionTime":"2025-09-29T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.287579 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.287734 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.287815 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.287924 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.287998 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:58Z","lastTransitionTime":"2025-09-29T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.390126 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.390156 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.390168 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.390179 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.390187 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:58Z","lastTransitionTime":"2025-09-29T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.492366 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.492414 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.492423 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.492434 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.492441 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:58Z","lastTransitionTime":"2025-09-29T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.594171 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.594205 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.594213 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.594226 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.594233 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:58Z","lastTransitionTime":"2025-09-29T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.695903 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.695928 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.695936 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.695946 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.695954 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:58Z","lastTransitionTime":"2025-09-29T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.797220 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.797242 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.797251 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.797260 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.797267 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:58Z","lastTransitionTime":"2025-09-29T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.898586 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.898631 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.898655 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.898666 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:58 crc kubenswrapper[4667]: I0929 17:10:58.898672 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:58Z","lastTransitionTime":"2025-09-29T17:10:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.000492 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.000521 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.000529 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.000540 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.000548 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.102124 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.102155 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.102191 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.102211 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.102221 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.203461 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.203505 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.203517 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.203534 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.203546 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.305003 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.305030 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.305040 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.305052 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.305061 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.406682 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.406786 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.406877 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.406946 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.407027 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.508701 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.508730 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.508739 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.508750 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.508757 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.611149 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.611178 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.611186 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.611198 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.611206 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.713203 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.713257 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.713266 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.713278 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.713286 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.814702 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:10:59 crc kubenswrapper[4667]: E0929 17:10:59.814798 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.814824 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.814824 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.814836 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:10:59 crc kubenswrapper[4667]: E0929 17:10:59.814941 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:10:59 crc kubenswrapper[4667]: E0929 17:10:59.815011 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.815023 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.815046 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.815055 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.815067 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: E0929 17:10:59.815065 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.815075 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.916440 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.916469 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.916477 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.916487 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:10:59 crc kubenswrapper[4667]: I0929 17:10:59.916495 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:10:59Z","lastTransitionTime":"2025-09-29T17:10:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.018785 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.018809 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.018817 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.018829 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.018857 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.120879 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.120908 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.120916 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.120927 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.120937 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.222446 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.222479 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.222486 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.222514 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.222522 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.324539 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.324689 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.324750 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.324823 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.324895 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.427145 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.427269 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.427343 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.427407 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.427474 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.528891 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.529061 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.529143 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.529206 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.529262 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.631373 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.631571 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.631657 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.631726 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.631795 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.733992 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.734304 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.734466 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.734611 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.734740 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.836215 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.836245 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.836254 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.836289 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.836297 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.938458 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.938558 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.938630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.938696 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:00 crc kubenswrapper[4667]: I0929 17:11:00.938750 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:00Z","lastTransitionTime":"2025-09-29T17:11:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.040819 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.040885 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.040897 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.040911 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.040921 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.143134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.143170 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.143180 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.143193 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.143203 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.244813 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.244868 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.244880 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.244893 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.244901 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.346762 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.346796 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.346806 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.346820 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.346830 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.448626 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.448660 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.448671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.448684 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.448693 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.550923 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.550964 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.550973 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.550986 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.550994 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.652554 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.652648 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.652659 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.652669 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.652677 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.753915 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.754248 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.754313 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.754380 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.754449 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.815306 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.815381 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:01 crc kubenswrapper[4667]: E0929 17:11:01.815458 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.815507 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.815516 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:01 crc kubenswrapper[4667]: E0929 17:11:01.815597 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:01 crc kubenswrapper[4667]: E0929 17:11:01.815732 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:01 crc kubenswrapper[4667]: E0929 17:11:01.815838 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.823227 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.832236 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.844762 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fc84a9a-2ed1-4304-b7e8-37d43fee7d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb7c287f124ce6b9696a6ae200279ba2e4509201c42176c5f1c21d78f948ea3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbfa2510696eb81866c2158c7eb7102a0c591f568ed85b48ba594d769ddd98d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d89da318419cde5f78e062ae14168f716c169fb0bb7a2a7468974021ff75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57073aea9b43bae01cf9e3f0e23390e2550bc3942f07cf94bdb9b92f12d4a9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6f02c341c7866bad9a9a485185b47b00d93f8d0b2f289d27f1ef851edbc18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fe7aa4a24221e184afdd808a3a04146b5fb2ef77c3720ca69a1358aff749618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe7aa4a24221e184afdd808a3a04146b5fb2ef77c3720ca69a1358aff749618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ed387e17159598d2c11525b5534226b01ac0f7beb08e43aed9d93024de185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b98ed387e17159598d2c11525b5534226b01ac0f7beb08e43aed9d93024de185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6ae461746f91f42609f4b85cf93ffcfd68aa9b865190a44af69532e600a55f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae461746f91f42609f4b85cf93ffcfd68aa9b865190a44af69532e600a55f46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.852369 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.855585 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.855654 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.855666 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.855677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.855686 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.861209 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.868821 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.876231 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.882533 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.889923 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.896626 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.905529 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.913117 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.919785 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.927689 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:37Z\\\",\\\"message\\\":\\\"2025-09-29T17:09:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96\\\\n2025-09-29T17:09:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96 to /host/opt/cni/bin/\\\\n2025-09-29T17:09:52Z [verbose] multus-daemon started\\\\n2025-09-29T17:09:52Z [verbose] Readiness Indicator file check\\\\n2025-09-29T17:10:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.939576 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"365] Adding new object: *v1.Pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422928 6733 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-h7cv4 in node crc\\\\nI0929 17:10:43.422932 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-h7cv4 after 0 failed attempt(s)\\\\nI0929 17:10:43.422935 6733 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422809 6733 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nI0929 17:10:43.422941 6733 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-l8rmj in node crc\\\\nI0929 17:10:43.422945 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj after 0 failed attempt(s)\\\\nI0929 17:10:43.422949 6733 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nF0929 17:10:43.422795 6733 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller init\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.949199 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.957105 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.957651 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.957671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.957680 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.957690 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.957697 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:01Z","lastTransitionTime":"2025-09-29T17:11:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:01 crc kubenswrapper[4667]: I0929 17:11:01.963093 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:01Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.059926 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.059957 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.059965 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.059975 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.059982 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.161242 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.161273 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.161282 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.161294 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.161302 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.263014 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.263038 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.263047 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.263057 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.263065 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.365134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.365227 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.365291 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.365359 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.365417 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.467229 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.467261 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.467270 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.467283 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.467292 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.569127 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.569227 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.569290 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.569348 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.569409 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.671538 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.671678 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.671743 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.671806 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.671876 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.773762 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.774118 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.774185 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.774253 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.774312 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.823563 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.876346 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.876384 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.876393 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.876404 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.876413 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.978006 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.978134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.978222 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.978306 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:02 crc kubenswrapper[4667]: I0929 17:11:02.978369 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:02Z","lastTransitionTime":"2025-09-29T17:11:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.079703 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.079728 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.079736 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.079746 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.079752 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.181601 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.181634 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.181642 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.181654 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.181663 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.283669 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.283695 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.283703 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.283714 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.283722 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.385346 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.385378 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.385387 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.385408 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.385420 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.487258 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.487374 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.487433 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.487492 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.487550 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.589036 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.589064 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.589072 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.589082 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.589089 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.690789 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.690817 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.690825 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.690834 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.690867 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.793130 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.793168 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.793177 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.793190 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.793200 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.814649 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.814698 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.814693 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.814710 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:03 crc kubenswrapper[4667]: E0929 17:11:03.814780 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:03 crc kubenswrapper[4667]: E0929 17:11:03.814901 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:03 crc kubenswrapper[4667]: E0929 17:11:03.815228 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:03 crc kubenswrapper[4667]: E0929 17:11:03.815300 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.895518 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.895555 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.895563 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.895575 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.895583 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.997692 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.997730 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.997740 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.997750 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:03 crc kubenswrapper[4667]: I0929 17:11:03.997758 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:03Z","lastTransitionTime":"2025-09-29T17:11:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.082340 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.082361 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.082369 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.082380 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.082389 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: E0929 17:11:04.091276 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.093633 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.093655 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.093663 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.093672 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.093679 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: E0929 17:11:04.101957 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.104092 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.104115 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.104139 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.104148 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.104155 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: E0929 17:11:04.111859 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.114028 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.114066 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.114074 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.114084 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.114092 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: E0929 17:11:04.121988 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.123970 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.123995 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.124003 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.124012 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.124018 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: E0929 17:11:04.131317 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:04Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:04 crc kubenswrapper[4667]: E0929 17:11:04.131425 4667 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.132405 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.132439 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.132448 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.132460 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.132486 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.233825 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.233862 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.233871 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.233880 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.233886 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.335430 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.335480 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.335490 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.335503 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.335513 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.436812 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.436871 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.436881 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.436890 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.436899 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.538987 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.539016 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.539024 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.539033 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.539058 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.640107 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.640146 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.640157 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.640167 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.640174 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.741988 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.742019 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.742027 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.742038 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.742046 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.843200 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.843228 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.843239 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.843248 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.843254 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.944960 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.944987 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.944995 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.945006 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:04 crc kubenswrapper[4667]: I0929 17:11:04.945013 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:04Z","lastTransitionTime":"2025-09-29T17:11:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.046578 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.046601 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.046609 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.046618 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.046625 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.147453 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.147480 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.147488 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.147497 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.147504 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.249295 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.249319 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.249326 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.249335 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.249342 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.351458 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.351504 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.351512 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.351525 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.351534 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.455104 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.455137 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.455147 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.455160 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.455171 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.556801 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.556835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.556862 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.556875 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.556884 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.658828 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.658908 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.658917 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.658930 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.658940 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.760913 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.760949 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.760958 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.760971 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.760983 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.815324 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.815359 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:05 crc kubenswrapper[4667]: E0929 17:11:05.815409 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.815324 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:05 crc kubenswrapper[4667]: E0929 17:11:05.815480 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:05 crc kubenswrapper[4667]: E0929 17:11:05.815516 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.815528 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:05 crc kubenswrapper[4667]: E0929 17:11:05.815641 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.862370 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.862396 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.862403 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.862413 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.862421 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.964387 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.964430 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.964438 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.964451 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:05 crc kubenswrapper[4667]: I0929 17:11:05.964459 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:05Z","lastTransitionTime":"2025-09-29T17:11:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.065914 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.065948 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.065956 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.065967 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.065975 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.167276 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.167305 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.167313 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.167324 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.167333 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.268857 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.268887 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.268908 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.268917 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.268924 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.370691 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.370729 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.370737 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.370751 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.370760 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.472501 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.472530 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.472537 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.472546 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.472555 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.574463 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.574491 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.574499 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.574510 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.574517 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.676230 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.676264 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.676273 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.676286 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.676295 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.778296 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.778325 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.778333 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.778345 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.778354 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.880144 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.880179 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.880187 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.880198 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.880212 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.981565 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.981592 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.981601 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.981611 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:06 crc kubenswrapper[4667]: I0929 17:11:06.981619 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:06Z","lastTransitionTime":"2025-09-29T17:11:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.082737 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.082770 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.082779 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.082790 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.082798 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.185192 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.185220 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.185228 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.185237 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.185245 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.286866 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.286903 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.286917 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.286928 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.286935 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.388821 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.388903 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.388914 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.388929 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.388939 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.490688 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.490715 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.490723 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.490748 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.490765 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.561266 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:07 crc kubenswrapper[4667]: E0929 17:11:07.561383 4667 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:11:07 crc kubenswrapper[4667]: E0929 17:11:07.561426 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs podName:d360e6c4-2b40-4214-bb7c-5d08038c1b62 nodeName:}" failed. No retries permitted until 2025-09-29 17:12:11.56141457 +0000 UTC m=+160.059261339 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs") pod "network-metrics-daemon-cl5p9" (UID: "d360e6c4-2b40-4214-bb7c-5d08038c1b62") : object "openshift-multus"/"metrics-daemon-secret" not registered Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.592653 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.592682 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.592690 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.592701 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.592710 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.694328 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.694357 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.694365 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.694376 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.694385 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.796207 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.796236 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.796244 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.796257 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.796265 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.815552 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.815604 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.815730 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.815785 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:07 crc kubenswrapper[4667]: E0929 17:11:07.815954 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:07 crc kubenswrapper[4667]: E0929 17:11:07.816052 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:07 crc kubenswrapper[4667]: E0929 17:11:07.816315 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:07 crc kubenswrapper[4667]: E0929 17:11:07.816369 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.816511 4667 scope.go:117] "RemoveContainer" containerID="d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415" Sep 29 17:11:07 crc kubenswrapper[4667]: E0929 17:11:07.816623 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.897254 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.897282 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.897290 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.897300 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.897308 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.999416 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.999437 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.999445 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.999454 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:07 crc kubenswrapper[4667]: I0929 17:11:07.999462 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:07Z","lastTransitionTime":"2025-09-29T17:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.101020 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.101049 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.101057 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.101066 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.101073 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:08Z","lastTransitionTime":"2025-09-29T17:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.202683 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.202720 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.202729 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.202743 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.202753 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:08Z","lastTransitionTime":"2025-09-29T17:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.304151 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.304182 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.304190 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.304199 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.304208 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:08Z","lastTransitionTime":"2025-09-29T17:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.405885 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.405921 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.405929 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.405942 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.405950 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:08Z","lastTransitionTime":"2025-09-29T17:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.506941 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.506963 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.506972 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.506983 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.506992 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:08Z","lastTransitionTime":"2025-09-29T17:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.609175 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.609225 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.609250 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.609263 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.609273 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:08Z","lastTransitionTime":"2025-09-29T17:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.710685 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.710715 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.710723 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.710735 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.710744 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:08Z","lastTransitionTime":"2025-09-29T17:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.812918 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.812947 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.812956 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.812965 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.812972 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:08Z","lastTransitionTime":"2025-09-29T17:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.914171 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.914193 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.914201 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.914210 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:08 crc kubenswrapper[4667]: I0929 17:11:08.914218 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:08Z","lastTransitionTime":"2025-09-29T17:11:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.016020 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.016073 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.016086 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.016102 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.016115 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.118028 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.118046 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.118053 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.118061 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.118068 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.219312 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.219580 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.219647 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.219714 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.219782 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.321714 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.321960 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.322076 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.322226 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.322367 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.424062 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.424088 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.424097 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.424107 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.424114 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.525757 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.526077 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.526138 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.526208 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.526276 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.628308 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.628334 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.628341 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.628350 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.628357 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.730515 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.730552 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.730560 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.730574 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.730582 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.814942 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.815012 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:09 crc kubenswrapper[4667]: E0929 17:11:09.815105 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.815149 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.815174 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:09 crc kubenswrapper[4667]: E0929 17:11:09.815218 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:09 crc kubenswrapper[4667]: E0929 17:11:09.815282 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:09 crc kubenswrapper[4667]: E0929 17:11:09.815338 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.831677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.831705 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.831713 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.831722 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.831729 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.933534 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.933570 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.933581 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.933595 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:09 crc kubenswrapper[4667]: I0929 17:11:09.933604 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:09Z","lastTransitionTime":"2025-09-29T17:11:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.035197 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.035223 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.035231 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.035243 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.035251 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.136564 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.136592 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.136600 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.136611 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.136620 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.238518 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.238551 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.238559 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.238571 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.238580 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.340445 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.340480 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.340490 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.340501 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.340511 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.442061 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.442092 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.442100 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.442111 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.442118 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.543914 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.543958 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.543968 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.543982 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.543992 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.645516 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.645568 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.645578 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.645590 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.645597 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.747440 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.747492 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.747501 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.747512 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.747520 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.849566 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.849612 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.849624 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.849635 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.849643 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.951880 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.951906 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.951913 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.951924 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:10 crc kubenswrapper[4667]: I0929 17:11:10.951932 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:10Z","lastTransitionTime":"2025-09-29T17:11:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.053449 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.053482 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.053491 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.053504 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.053514 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.155472 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.155517 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.155526 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.155537 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.155545 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.256795 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.256826 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.256873 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.256883 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.256892 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.358643 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.358692 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.358703 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.358714 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.358723 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.460131 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.460158 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.460166 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.460178 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.460186 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.561164 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.561187 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.561194 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.561204 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.561211 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.663031 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.663079 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.663088 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.663098 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.663104 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.764562 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.764589 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.764598 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.764609 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.764619 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.815064 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.815140 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.815225 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:11 crc kubenswrapper[4667]: E0929 17:11:11.815226 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.815251 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:11 crc kubenswrapper[4667]: E0929 17:11:11.815379 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:11 crc kubenswrapper[4667]: E0929 17:11:11.815452 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:11 crc kubenswrapper[4667]: E0929 17:11:11.815537 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.824234 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cec7feac5fd77f5c630aa95a4cb2b4a669af22a58fcc7023f36ffd9b1ec92fa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0d3f149f7919d72620c9081d63e7219596e6459d7c7859233d7d3ee9f91a63e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.831227 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d360e6c4-2b40-4214-bb7c-5d08038c1b62\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6ng6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:03Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-cl5p9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.837468 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-h7cv4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"812bd13c-b2d7-4e1a-a226-5794831f8c6a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1aef20af98b05a4c3ba887d6acf081533a9bf16ed223e0add40eee0302d4c340\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6g4c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:53Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-h7cv4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.845470 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"daa52c6c-9504-473e-b842-86e41197131a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578396b629294f9001814df52d1e980151dbfbc35ad7b6ae8a96e9355b8d2189\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15d7ff64f581baa39313ce5e7ec26aeaa4d19895a546514b29f2949eb86d190c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1050d705fb9ef3563b24cb623ef9264453ef81118d798f3a3f7e23919320c01\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26bb8ed0e2b40aef592ee0a040fe2bbaf5fb8bbfb3123584cc260395f024e43d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.857737 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6fc84a9a-2ed1-4304-b7e8-37d43fee7d7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb7c287f124ce6b9696a6ae200279ba2e4509201c42176c5f1c21d78f948ea3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6bbfa2510696eb81866c2158c7eb7102a0c591f568ed85b48ba594d769ddd98d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9d89da318419cde5f78e062ae14168f716c169fb0bb7a2a7468974021ff75b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://57073aea9b43bae01cf9e3f0e23390e2550bc3942f07cf94bdb9b92f12d4a9ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://43d6f02c341c7866bad9a9a485185b47b00d93f8d0b2f289d27f1ef851edbc18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fe7aa4a24221e184afdd808a3a04146b5fb2ef77c3720ca69a1358aff749618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fe7aa4a24221e184afdd808a3a04146b5fb2ef77c3720ca69a1358aff749618\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b98ed387e17159598d2c11525b5534226b01ac0f7beb08e43aed9d93024de185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b98ed387e17159598d2c11525b5534226b01ac0f7beb08e43aed9d93024de185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6ae461746f91f42609f4b85cf93ffcfd68aa9b865190a44af69532e600a55f46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6ae461746f91f42609f4b85cf93ffcfd68aa9b865190a44af69532e600a55f46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.866302 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.866330 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.866410 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.866419 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.866430 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.866439 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.875090 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5mghc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61c9ad8c-4cc7-4eeb-b985-3ef2a0893fc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4f40be4ab3c9354ff6f964de55f94f303a49ab19a073f7f8ba56189c784bdf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a4a1e214a5ec555cbeaa2d9ed520ada4392b4337b7859ee69b3bd17d75fca7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f75591aa453263ce4a026e40a3c6cd7605dc9c81b4dd2149e6033b504ea64f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e379b137b6a43e7b9844e012e280530f174fea0a6e0dd6cc09b1e2d9f8eaccc5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce70e588aab733984e4e07819909a35c4918dde201fb14a4a2f53e6fe0754390\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9d1363b6b8a4876b569b9f4ac1dcaf8a93d4c5500afd8d25bfe92b49601d9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bce674e1692be2f088950eeff337ef16f8bd9720132939bcc0632738a0da08a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dhgkq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5mghc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.883174 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.890411 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lbbgf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7192692e-a16e-46c5-9097-0e15418054a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5b2646edd7a0699a9e57b299e3bc1e6311cab09554b37b63194e022aa09ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9hv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:49Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lbbgf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.897353 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://81715d1361b8dae679fad7886e2bd5d7d3f9c20f8b6daba539371bcb462b63f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.904110 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2a4667c4-91a0-4513-b149-8b5afc40c405\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd854c6da5f63ee3b729ecfb5bfd031d7bb03e24450e2ba4289c7b09eef60c48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b12d26d0abea12c956c81fd62f6471ca2c61b3c6cd403e790b0bae96839a59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0486d4ea6799358d7d2a4b32d7bc8a9ed914770ec65b9e7688c33727eec1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52018ced99f1fd84bbc65a05c03b2fcf8215849757ca230c59360a3e325b9e36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.910102 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0f826469-7610-4cca-97e6-5a5bd0a6f692\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e1c15dc05de758c79f9d53e0d55d88187ddbf28778b9d70a52cd19d31a3d47a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7e08e45811bd7a6442cc15a3212ba001bc4ef2df1ddda320c9a139f7fd01fdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b7e08e45811bd7a6442cc15a3212ba001bc4ef2df1ddda320c9a139f7fd01fdd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.920717 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3bb654e3-75d8-4128-9262-2ad0cd2e612c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e69f17a8141a4cf7e5b64a0b09b39bfa79bb3d8db2601f9f870bc13b0155eed0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c7881f656d4cacee987bf047291cf68b583d50cee8685ac110f6085ade0462\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://932721b11995b71931f27d185113f31efe40053854aa368d3efa23fd1931c5e1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://965358d7680f57b7dd08b422b837a20c0dc13978d64b4e666a7a393bc8d716fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e2b42ed3b4343a0c31a3ad1e011f7e628f7bf1efc8cfb3d4c0db55723847bf92\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0929 17:09:43.728385 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0929 17:09:43.729614 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2942105101/tls.crt::/tmp/serving-cert-2942105101/tls.key\\\\\\\"\\\\nI0929 17:09:49.430938 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0929 17:09:49.434000 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0929 17:09:49.434053 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0929 17:09:49.434083 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0929 17:09:49.434092 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0929 17:09:49.440317 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0929 17:09:49.440340 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0929 17:09:49.440348 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0929 17:09:49.440351 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0929 17:09:49.440354 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0929 17:09:49.440356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0929 17:09:49.440398 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0929 17:09:49.442478 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4382b96c1aab7af6f7697c0a29f5bbfbe5b2707db475a2a160bd7fa8904ff576\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:33Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2bfa8488a1e2e88acf48502e8359cee583e60a35f002fb0965cf07af8d4ce767\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:31Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.928763 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:50Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8559bc94543611a9e97cb3595404b2758ac7c9463cd9293a4eca6357f6417003\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.936273 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:49Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.943650 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28fa0016-3e75-4704-8b60-30ee9e576d59\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85747d8adb49a3d535d023ff7bdf5e38b18ef81886d79b45db4c0d3a8533c860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wh4nc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-l8rmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.951562 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2bf6m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"788f770a-3181-4b66-981c-90ffb7fc49c0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:37Z\\\",\\\"message\\\":\\\"2025-09-29T17:09:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96\\\\n2025-09-29T17:09:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_1f0481b2-b648-4de6-bedc-4cbf05b43f96 to /host/opt/cni/bin/\\\\n2025-09-29T17:09:52Z [verbose] multus-daemon started\\\\n2025-09-29T17:09:52Z [verbose] Readiness Indicator file check\\\\n2025-09-29T17:10:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q54hx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2bf6m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.962967 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6250a9ff-80f5-44d8-90f6-40e77106af6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-09-29T17:10:43Z\\\",\\\"message\\\":\\\"365] Adding new object: *v1.Pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422928 6733 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-h7cv4 in node crc\\\\nI0929 17:10:43.422932 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-h7cv4 after 0 failed attempt(s)\\\\nI0929 17:10:43.422935 6733 default_network_controller.go:776] Recording success event on pod openshift-image-registry/node-ca-h7cv4\\\\nI0929 17:10:43.422809 6733 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nI0929 17:10:43.422941 6733 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-l8rmj in node crc\\\\nI0929 17:10:43.422945 6733 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-l8rmj after 0 failed attempt(s)\\\\nI0929 17:10:43.422949 6733 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-l8rmj\\\\nF0929 17:10:43.422795 6733 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller init\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-09-29T17:10:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:09:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-09-29T17:09:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-09-29T17:09:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bqs9s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:09:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qjsnt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.968340 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.968367 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.968375 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.968387 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.968395 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:11Z","lastTransitionTime":"2025-09-29T17:11:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:11 crc kubenswrapper[4667]: I0929 17:11:11.969980 4667 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64e4a86f-c3e1-4a00-b3c4-1d1294635c9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-09-29T17:10:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c859aa7888a138ff6a3490333956e48a3e50bb5b2c02e81807fcb8400ba41d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e767fc5655e7d4ed80ea419f6e8e9688a32bdc3347ab15085bfde8081fe4117e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-09-29T17:10:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jmmr8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-09-29T17:10:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-ljv29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:11Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.069637 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.069668 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.069677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.069689 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.069697 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.170741 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.170775 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.170785 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.170797 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.170807 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.272639 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.272687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.272696 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.272708 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.272717 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.374674 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.374704 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.374714 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.374724 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.374732 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.476108 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.476134 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.476142 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.476152 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.476160 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.577949 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.577976 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.577987 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.578000 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.578007 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.680071 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.680119 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.680127 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.680136 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.680145 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.781447 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.781471 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.781479 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.781489 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.781496 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.882751 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.882777 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.882784 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.882795 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.882802 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.984634 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.984689 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.984700 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.984712 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:12 crc kubenswrapper[4667]: I0929 17:11:12.984720 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:12Z","lastTransitionTime":"2025-09-29T17:11:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.086763 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.086780 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.086788 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.086799 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.086816 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:13Z","lastTransitionTime":"2025-09-29T17:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.187798 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.187875 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.187885 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.187898 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.187906 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:13Z","lastTransitionTime":"2025-09-29T17:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.289181 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.289215 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.289226 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.289265 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.289273 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:13Z","lastTransitionTime":"2025-09-29T17:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.390674 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.390709 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.390719 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.390732 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.390740 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:13Z","lastTransitionTime":"2025-09-29T17:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.492443 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.492475 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.492484 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.492499 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.492509 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:13Z","lastTransitionTime":"2025-09-29T17:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.593708 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.593730 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.593739 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.593748 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.593756 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:13Z","lastTransitionTime":"2025-09-29T17:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.695533 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.695560 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.695568 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.695576 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.695583 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:13Z","lastTransitionTime":"2025-09-29T17:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.797619 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.797653 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.797662 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.797673 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.797682 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:13Z","lastTransitionTime":"2025-09-29T17:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.815167 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.815192 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:13 crc kubenswrapper[4667]: E0929 17:11:13.815244 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.815280 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.815296 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:13 crc kubenswrapper[4667]: E0929 17:11:13.815400 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:13 crc kubenswrapper[4667]: E0929 17:11:13.815438 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:13 crc kubenswrapper[4667]: E0929 17:11:13.815464 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.899532 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.899575 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.899583 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.899594 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:13 crc kubenswrapper[4667]: I0929 17:11:13.899602 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:13Z","lastTransitionTime":"2025-09-29T17:11:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.001716 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.001752 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.001761 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.001773 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.001779 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.103978 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.104013 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.104021 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.104035 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.104043 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.206130 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.206165 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.206174 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.206186 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.206194 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.308116 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.308157 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.308167 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.308178 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.308186 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.404149 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.404183 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.404192 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.404206 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.404214 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: E0929 17:11:14.412718 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:14Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.415233 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.415260 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.415268 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.415279 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.415286 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: E0929 17:11:14.422815 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:14Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.424774 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.424805 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.424815 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.424827 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.424834 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: E0929 17:11:14.433496 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:14Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.435552 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.435576 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.435585 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.435595 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.435603 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: E0929 17:11:14.443048 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:14Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.445025 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.445050 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.445059 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.445068 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.445075 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: E0929 17:11:14.452575 4667 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404548Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865348Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-09-29T17:11:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"f7158ada-47b4-429a-bd74-dd92a5b97fd6\\\",\\\"systemUUID\\\":\\\"d5a94666-8121-4bfb-8540-72964a1282ac\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-09-29T17:11:14Z is after 2025-08-24T17:21:41Z" Sep 29 17:11:14 crc kubenswrapper[4667]: E0929 17:11:14.452677 4667 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.453648 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.453677 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.453687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.453698 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.453707 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.555381 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.555437 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.555446 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.555458 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.555467 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.657155 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.657179 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.657187 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.657198 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.657205 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.759128 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.759163 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.759171 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.759199 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.759209 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.861269 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.861318 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.861329 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.861344 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.861353 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.963443 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.963477 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.963486 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.963515 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:14 crc kubenswrapper[4667]: I0929 17:11:14.963523 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:14Z","lastTransitionTime":"2025-09-29T17:11:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.066048 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.066081 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.066090 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.066100 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.066106 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.168301 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.168369 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.168382 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.168403 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.168416 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.270050 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.270095 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.270104 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.270119 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.270130 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.372310 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.372340 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.372350 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.372360 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.372368 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.473714 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.473761 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.473769 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.473794 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.473802 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.575649 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.575678 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.575687 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.575697 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.575705 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.678034 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.678090 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.678103 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.678124 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.678134 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.779564 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.779613 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.779623 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.779638 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.779647 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.815491 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.815516 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.815546 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:15 crc kubenswrapper[4667]: E0929 17:11:15.815664 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.815703 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:15 crc kubenswrapper[4667]: E0929 17:11:15.815754 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:15 crc kubenswrapper[4667]: E0929 17:11:15.815832 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:15 crc kubenswrapper[4667]: E0929 17:11:15.815891 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.881464 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.881491 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.881499 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.881509 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.881515 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.982982 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.983008 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.983017 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.983027 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:15 crc kubenswrapper[4667]: I0929 17:11:15.983034 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:15Z","lastTransitionTime":"2025-09-29T17:11:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.085053 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.085291 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.085301 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.085311 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.085318 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:16Z","lastTransitionTime":"2025-09-29T17:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.186576 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.186611 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.186621 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.186631 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.186637 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:16Z","lastTransitionTime":"2025-09-29T17:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.288025 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.288063 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.288073 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.288083 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.288092 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:16Z","lastTransitionTime":"2025-09-29T17:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.389786 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.389809 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.389817 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.389826 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.389833 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:16Z","lastTransitionTime":"2025-09-29T17:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.491588 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.491619 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.491628 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.491639 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.491646 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:16Z","lastTransitionTime":"2025-09-29T17:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.592990 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.593017 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.593025 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.593035 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.593042 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:16Z","lastTransitionTime":"2025-09-29T17:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.694640 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.694678 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.694690 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.694700 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.694706 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:16Z","lastTransitionTime":"2025-09-29T17:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.796704 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.796737 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.796746 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.796757 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.796774 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:16Z","lastTransitionTime":"2025-09-29T17:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.898743 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.898772 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.898782 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.898791 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:16 crc kubenswrapper[4667]: I0929 17:11:16.898798 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:16Z","lastTransitionTime":"2025-09-29T17:11:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.000371 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.000391 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.000399 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.000409 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.000416 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.102287 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.102333 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.102344 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.102358 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.102369 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.204253 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.204312 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.204330 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.204354 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.204368 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.306715 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.306755 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.306775 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.306796 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.306806 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.408635 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.408671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.408683 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.408697 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.408706 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.510298 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.510326 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.510345 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.510360 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.510368 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.612809 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.612858 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.612869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.612882 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.612891 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.714797 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.714824 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.714832 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.714865 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.714876 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.814605 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.814654 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:17 crc kubenswrapper[4667]: E0929 17:11:17.814714 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:17 crc kubenswrapper[4667]: E0929 17:11:17.814777 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.814829 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:17 crc kubenswrapper[4667]: E0929 17:11:17.814903 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.815108 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:17 crc kubenswrapper[4667]: E0929 17:11:17.815282 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.816173 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.816203 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.816218 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.816228 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.816237 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.918279 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.918309 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.918319 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.918330 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:17 crc kubenswrapper[4667]: I0929 17:11:17.918339 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:17Z","lastTransitionTime":"2025-09-29T17:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.019381 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.019506 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.019595 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.019666 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.019730 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.122511 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.122534 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.122542 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.122553 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.122564 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.224025 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.224053 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.224061 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.224071 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.224079 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.325389 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.325416 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.325426 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.325437 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.325450 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.427000 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.427022 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.427030 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.427046 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.427055 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.529001 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.529050 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.529067 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.529082 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.529093 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.630333 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.630361 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.630368 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.630380 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.630387 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.731958 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.731984 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.731993 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.732006 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.732013 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.834013 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.834038 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.834046 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.834056 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.834062 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.935296 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.935321 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.935329 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.935337 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:18 crc kubenswrapper[4667]: I0929 17:11:18.935344 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:18Z","lastTransitionTime":"2025-09-29T17:11:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.036886 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.036919 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.036929 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.036941 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.036950 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.138622 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.138649 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.138659 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.138668 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.138675 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.240561 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.240588 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.240595 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.240606 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.240613 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.342183 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.342211 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.342220 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.342229 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.342235 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.444279 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.444315 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.444324 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.444336 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.444345 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.546232 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.546255 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.546263 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.546272 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.546279 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.647971 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.648018 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.648026 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.648041 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.648051 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.749799 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.749819 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.749827 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.749835 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.749862 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.814806 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:19 crc kubenswrapper[4667]: E0929 17:11:19.814908 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.814946 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.815081 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.815193 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:19 crc kubenswrapper[4667]: E0929 17:11:19.815270 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:19 crc kubenswrapper[4667]: E0929 17:11:19.815449 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:19 crc kubenswrapper[4667]: E0929 17:11:19.815494 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.851007 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.851036 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.851044 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.851053 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.851060 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.952944 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.953083 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.953149 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.953220 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:19 crc kubenswrapper[4667]: I0929 17:11:19.953281 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:19Z","lastTransitionTime":"2025-09-29T17:11:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.054916 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.054962 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.054975 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.054997 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.055011 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.156572 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.156600 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.156609 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.156618 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.156624 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.258786 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.258810 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.258818 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.258827 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.258834 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.360596 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.360617 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.360626 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.360635 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.360642 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.462463 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.462496 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.462504 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.462516 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.462525 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.563867 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.563890 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.563899 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.563909 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.563917 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.665732 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.665752 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.665759 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.665768 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.665774 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.766920 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.766952 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.766961 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.766975 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.766983 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.869091 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.869515 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.869599 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.869671 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.869746 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.971699 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.971742 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.971752 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.971763 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:20 crc kubenswrapper[4667]: I0929 17:11:20.971770 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:20Z","lastTransitionTime":"2025-09-29T17:11:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.073540 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.073565 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.073574 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.073584 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.073592 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.175019 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.175045 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.175053 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.175063 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.175071 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.276818 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.276946 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.277005 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.277085 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.277143 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.379243 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.379264 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.379271 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.379281 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.379289 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.481363 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.481389 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.481398 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.481409 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.481417 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.582804 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.582831 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.582856 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.582871 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.582879 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.684696 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.684759 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.684774 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.684790 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.684801 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.786110 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.786133 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.786140 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.786150 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.786157 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.814780 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.814813 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:21 crc kubenswrapper[4667]: E0929 17:11:21.814902 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.814909 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.814968 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:21 crc kubenswrapper[4667]: E0929 17:11:21.815020 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:21 crc kubenswrapper[4667]: E0929 17:11:21.815049 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:21 crc kubenswrapper[4667]: E0929 17:11:21.815204 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.852976 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-h7cv4" podStartSLOduration=91.852963309 podStartE2EDuration="1m31.852963309s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.842658843 +0000 UTC m=+110.340505611" watchObservedRunningTime="2025-09-29 17:11:21.852963309 +0000 UTC m=+110.350810078" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.862250 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-5mghc" podStartSLOduration=91.862229831 podStartE2EDuration="1m31.862229831s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.853089576 +0000 UTC m=+110.350936345" watchObservedRunningTime="2025-09-29 17:11:21.862229831 +0000 UTC m=+110.360076610" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.862382 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=89.86237896 podStartE2EDuration="1m29.86237896s" podCreationTimestamp="2025-09-29 17:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.861789956 +0000 UTC m=+110.359636726" watchObservedRunningTime="2025-09-29 17:11:21.86237896 +0000 UTC m=+110.360225730" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.877347 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=33.877336861 podStartE2EDuration="33.877336861s" podCreationTimestamp="2025-09-29 17:10:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.876870116 +0000 UTC m=+110.374716885" watchObservedRunningTime="2025-09-29 17:11:21.877336861 +0000 UTC m=+110.375183631" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.887027 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.887199 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.887277 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.887366 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.887439 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.916690 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-lbbgf" podStartSLOduration=92.916672993 podStartE2EDuration="1m32.916672993s" podCreationTimestamp="2025-09-29 17:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.916652164 +0000 UTC m=+110.414498933" watchObservedRunningTime="2025-09-29 17:11:21.916672993 +0000 UTC m=+110.414519761" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.944373 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=57.944359926 podStartE2EDuration="57.944359926s" podCreationTimestamp="2025-09-29 17:10:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.937133659 +0000 UTC m=+110.434980428" watchObservedRunningTime="2025-09-29 17:11:21.944359926 +0000 UTC m=+110.442206695" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.944600 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=19.944596911 podStartE2EDuration="19.944596911s" podCreationTimestamp="2025-09-29 17:11:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.944307238 +0000 UTC m=+110.442154006" watchObservedRunningTime="2025-09-29 17:11:21.944596911 +0000 UTC m=+110.442443680" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.962435 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=92.96242573399999 podStartE2EDuration="1m32.962425734s" podCreationTimestamp="2025-09-29 17:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.955135757 +0000 UTC m=+110.452982526" watchObservedRunningTime="2025-09-29 17:11:21.962425734 +0000 UTC m=+110.460272503" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.962835 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-ljv29" podStartSLOduration=90.962831323 podStartE2EDuration="1m30.962831323s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.962118397 +0000 UTC m=+110.459965167" watchObservedRunningTime="2025-09-29 17:11:21.962831323 +0000 UTC m=+110.460678092" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.970024 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podStartSLOduration=91.970008228 podStartE2EDuration="1m31.970008228s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.969688929 +0000 UTC m=+110.467535698" watchObservedRunningTime="2025-09-29 17:11:21.970008228 +0000 UTC m=+110.467854997" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.978452 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2bf6m" podStartSLOduration=91.978443641 podStartE2EDuration="1m31.978443641s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:21.978060994 +0000 UTC m=+110.475907762" watchObservedRunningTime="2025-09-29 17:11:21.978443641 +0000 UTC m=+110.476290410" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.989801 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.989828 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.989854 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.989865 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:21 crc kubenswrapper[4667]: I0929 17:11:21.989875 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:21Z","lastTransitionTime":"2025-09-29T17:11:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.091494 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.091522 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.091531 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.091545 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.091555 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:22Z","lastTransitionTime":"2025-09-29T17:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.192553 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.192583 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.192592 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.192605 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.192614 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:22Z","lastTransitionTime":"2025-09-29T17:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.294478 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.294509 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.294518 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.294530 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.294546 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:22Z","lastTransitionTime":"2025-09-29T17:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.396136 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.396164 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.396172 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.396183 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.396190 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:22Z","lastTransitionTime":"2025-09-29T17:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.497804 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.497834 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.497857 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.497869 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.497877 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:22Z","lastTransitionTime":"2025-09-29T17:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.599343 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.599378 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.599388 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.599398 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.599404 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:22Z","lastTransitionTime":"2025-09-29T17:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.700897 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.700947 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.700956 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.700968 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.700976 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:22Z","lastTransitionTime":"2025-09-29T17:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.802101 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.802146 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.802158 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.802178 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.802187 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:22Z","lastTransitionTime":"2025-09-29T17:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.816024 4667 scope.go:117] "RemoveContainer" containerID="d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415" Sep 29 17:11:22 crc kubenswrapper[4667]: E0929 17:11:22.816138 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qjsnt_openshift-ovn-kubernetes(6250a9ff-80f5-44d8-90f6-40e77106af6c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.903892 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.903924 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.903953 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.903967 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:22 crc kubenswrapper[4667]: I0929 17:11:22.903976 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:22Z","lastTransitionTime":"2025-09-29T17:11:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.005034 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.005060 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.005067 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.005078 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.005085 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.106889 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.107085 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.107094 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.107106 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.107114 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.208709 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.208755 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.208765 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.208776 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.208784 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.310792 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.310819 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.310827 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.310838 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.310861 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.412661 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.412708 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.412717 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.412729 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.412737 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.514730 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.514761 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.514769 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.514780 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.514788 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.616371 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.616418 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.616426 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.616438 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.616445 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.718092 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.718118 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.718126 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.718137 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.718145 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.815605 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:23 crc kubenswrapper[4667]: E0929 17:11:23.815733 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.815781 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.815820 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.815838 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:23 crc kubenswrapper[4667]: E0929 17:11:23.815914 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:23 crc kubenswrapper[4667]: E0929 17:11:23.815991 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:23 crc kubenswrapper[4667]: E0929 17:11:23.816047 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.819411 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.819436 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.819444 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.819453 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.819461 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.921587 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.921622 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.921630 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.921643 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:23 crc kubenswrapper[4667]: I0929 17:11:23.921651 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:23Z","lastTransitionTime":"2025-09-29T17:11:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.023159 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.023198 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.023206 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.023216 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.023224 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:24Z","lastTransitionTime":"2025-09-29T17:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.124943 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.124969 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.124978 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.124988 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.125015 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:24Z","lastTransitionTime":"2025-09-29T17:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.178950 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/1.log" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.179346 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/0.log" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.179443 4667 generic.go:334] "Generic (PLEG): container finished" podID="788f770a-3181-4b66-981c-90ffb7fc49c0" containerID="150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033" exitCode=1 Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.179526 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2bf6m" event={"ID":"788f770a-3181-4b66-981c-90ffb7fc49c0","Type":"ContainerDied","Data":"150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.179569 4667 scope.go:117] "RemoveContainer" containerID="1088797f95799f9fc52d9d5fba2468fded9606c022d3749db47ce6874ec74c9c" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.179813 4667 scope.go:117] "RemoveContainer" containerID="150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033" Sep 29 17:11:24 crc kubenswrapper[4667]: E0929 17:11:24.179957 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-2bf6m_openshift-multus(788f770a-3181-4b66-981c-90ffb7fc49c0)\"" pod="openshift-multus/multus-2bf6m" podUID="788f770a-3181-4b66-981c-90ffb7fc49c0" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.226922 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.226966 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.226975 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.226986 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.226993 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:24Z","lastTransitionTime":"2025-09-29T17:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.328727 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.328760 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.328768 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.328781 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.328791 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:24Z","lastTransitionTime":"2025-09-29T17:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.430661 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.430708 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.430718 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.430732 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.430742 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:24Z","lastTransitionTime":"2025-09-29T17:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.531927 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.531952 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.531959 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.531967 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.531976 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:24Z","lastTransitionTime":"2025-09-29T17:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.633871 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.633909 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.633920 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.633933 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.633941 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:24Z","lastTransitionTime":"2025-09-29T17:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.735815 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.735896 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.735906 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.735918 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.735927 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:24Z","lastTransitionTime":"2025-09-29T17:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.791518 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.791547 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.791573 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.791584 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.791593 4667 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-29T17:11:24Z","lastTransitionTime":"2025-09-29T17:11:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.818726 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn"] Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.819043 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.820972 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.821000 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.821159 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.821225 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.896786 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbe01aed-0aba-4b4f-9308-479f02c48a78-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.896815 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbe01aed-0aba-4b4f-9308-479f02c48a78-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.896875 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cbe01aed-0aba-4b4f-9308-479f02c48a78-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.896921 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbe01aed-0aba-4b4f-9308-479f02c48a78-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.896945 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cbe01aed-0aba-4b4f-9308-479f02c48a78-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.997745 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cbe01aed-0aba-4b4f-9308-479f02c48a78-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.997769 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbe01aed-0aba-4b4f-9308-479f02c48a78-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.997787 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cbe01aed-0aba-4b4f-9308-479f02c48a78-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.997818 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbe01aed-0aba-4b4f-9308-479f02c48a78-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.997833 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbe01aed-0aba-4b4f-9308-479f02c48a78-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.997834 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cbe01aed-0aba-4b4f-9308-479f02c48a78-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.997905 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cbe01aed-0aba-4b4f-9308-479f02c48a78-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:24 crc kubenswrapper[4667]: I0929 17:11:24.998577 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cbe01aed-0aba-4b4f-9308-479f02c48a78-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:25 crc kubenswrapper[4667]: I0929 17:11:25.002066 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbe01aed-0aba-4b4f-9308-479f02c48a78-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:25 crc kubenswrapper[4667]: I0929 17:11:25.009905 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbe01aed-0aba-4b4f-9308-479f02c48a78-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6plgn\" (UID: \"cbe01aed-0aba-4b4f-9308-479f02c48a78\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:25 crc kubenswrapper[4667]: I0929 17:11:25.128211 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" Sep 29 17:11:25 crc kubenswrapper[4667]: I0929 17:11:25.183323 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" event={"ID":"cbe01aed-0aba-4b4f-9308-479f02c48a78","Type":"ContainerStarted","Data":"52524dcfa9bf4cadd6b6c4425d2712e549b3073f959a2e71d25556e582d25107"} Sep 29 17:11:25 crc kubenswrapper[4667]: I0929 17:11:25.184472 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/1.log" Sep 29 17:11:25 crc kubenswrapper[4667]: I0929 17:11:25.815150 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:25 crc kubenswrapper[4667]: I0929 17:11:25.815167 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:25 crc kubenswrapper[4667]: E0929 17:11:25.815794 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:25 crc kubenswrapper[4667]: I0929 17:11:25.815219 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:25 crc kubenswrapper[4667]: E0929 17:11:25.815908 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:25 crc kubenswrapper[4667]: I0929 17:11:25.815171 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:25 crc kubenswrapper[4667]: E0929 17:11:25.815988 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:25 crc kubenswrapper[4667]: E0929 17:11:25.815701 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:26 crc kubenswrapper[4667]: I0929 17:11:26.187311 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" event={"ID":"cbe01aed-0aba-4b4f-9308-479f02c48a78","Type":"ContainerStarted","Data":"c40c050fbd6ca02a60e43abebd6023a7afddd2c3b0c320c140111c5e8e5440ee"} Sep 29 17:11:26 crc kubenswrapper[4667]: I0929 17:11:26.197213 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6plgn" podStartSLOduration=96.197201878 podStartE2EDuration="1m36.197201878s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:26.196597726 +0000 UTC m=+114.694444495" watchObservedRunningTime="2025-09-29 17:11:26.197201878 +0000 UTC m=+114.695048648" Sep 29 17:11:27 crc kubenswrapper[4667]: I0929 17:11:27.815162 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:27 crc kubenswrapper[4667]: I0929 17:11:27.815238 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:27 crc kubenswrapper[4667]: E0929 17:11:27.815254 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:27 crc kubenswrapper[4667]: I0929 17:11:27.815262 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:27 crc kubenswrapper[4667]: I0929 17:11:27.815168 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:27 crc kubenswrapper[4667]: E0929 17:11:27.815350 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:27 crc kubenswrapper[4667]: E0929 17:11:27.815386 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:27 crc kubenswrapper[4667]: E0929 17:11:27.815419 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:29 crc kubenswrapper[4667]: I0929 17:11:29.815246 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:29 crc kubenswrapper[4667]: I0929 17:11:29.815274 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:29 crc kubenswrapper[4667]: E0929 17:11:29.815362 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:29 crc kubenswrapper[4667]: I0929 17:11:29.815371 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:29 crc kubenswrapper[4667]: I0929 17:11:29.815400 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:29 crc kubenswrapper[4667]: E0929 17:11:29.815483 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:29 crc kubenswrapper[4667]: E0929 17:11:29.815539 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:29 crc kubenswrapper[4667]: E0929 17:11:29.815598 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:31 crc kubenswrapper[4667]: I0929 17:11:31.815289 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:31 crc kubenswrapper[4667]: I0929 17:11:31.815361 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:31 crc kubenswrapper[4667]: I0929 17:11:31.815384 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:31 crc kubenswrapper[4667]: E0929 17:11:31.816197 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:31 crc kubenswrapper[4667]: I0929 17:11:31.816221 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:31 crc kubenswrapper[4667]: E0929 17:11:31.816313 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:31 crc kubenswrapper[4667]: E0929 17:11:31.816331 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:31 crc kubenswrapper[4667]: E0929 17:11:31.816357 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:31 crc kubenswrapper[4667]: E0929 17:11:31.855816 4667 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Sep 29 17:11:31 crc kubenswrapper[4667]: E0929 17:11:31.883625 4667 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:11:33 crc kubenswrapper[4667]: I0929 17:11:33.815593 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:33 crc kubenswrapper[4667]: I0929 17:11:33.815674 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:33 crc kubenswrapper[4667]: E0929 17:11:33.815703 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:33 crc kubenswrapper[4667]: I0929 17:11:33.815719 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:33 crc kubenswrapper[4667]: E0929 17:11:33.815791 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:33 crc kubenswrapper[4667]: I0929 17:11:33.815857 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:33 crc kubenswrapper[4667]: E0929 17:11:33.815916 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:33 crc kubenswrapper[4667]: E0929 17:11:33.815957 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:35 crc kubenswrapper[4667]: I0929 17:11:35.814771 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:35 crc kubenswrapper[4667]: I0929 17:11:35.814839 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:35 crc kubenswrapper[4667]: E0929 17:11:35.814896 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:35 crc kubenswrapper[4667]: I0929 17:11:35.814928 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:35 crc kubenswrapper[4667]: I0929 17:11:35.814948 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:35 crc kubenswrapper[4667]: E0929 17:11:35.814971 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:35 crc kubenswrapper[4667]: E0929 17:11:35.815022 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:35 crc kubenswrapper[4667]: E0929 17:11:35.815080 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:36 crc kubenswrapper[4667]: I0929 17:11:36.815164 4667 scope.go:117] "RemoveContainer" containerID="150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033" Sep 29 17:11:36 crc kubenswrapper[4667]: E0929 17:11:36.884727 4667 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:11:37 crc kubenswrapper[4667]: I0929 17:11:37.213968 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/1.log" Sep 29 17:11:37 crc kubenswrapper[4667]: I0929 17:11:37.214121 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2bf6m" event={"ID":"788f770a-3181-4b66-981c-90ffb7fc49c0","Type":"ContainerStarted","Data":"f8a3403dd025716fcd1c6fdc3bd819d34cd5e3d986a182788c97a3df512c1dd8"} Sep 29 17:11:37 crc kubenswrapper[4667]: I0929 17:11:37.815048 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:37 crc kubenswrapper[4667]: E0929 17:11:37.815298 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:37 crc kubenswrapper[4667]: I0929 17:11:37.815050 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:37 crc kubenswrapper[4667]: I0929 17:11:37.815445 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:37 crc kubenswrapper[4667]: I0929 17:11:37.815494 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:37 crc kubenswrapper[4667]: E0929 17:11:37.815578 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:37 crc kubenswrapper[4667]: E0929 17:11:37.815691 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:37 crc kubenswrapper[4667]: I0929 17:11:37.815734 4667 scope.go:117] "RemoveContainer" containerID="d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415" Sep 29 17:11:37 crc kubenswrapper[4667]: E0929 17:11:37.815798 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:38 crc kubenswrapper[4667]: I0929 17:11:38.218061 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/3.log" Sep 29 17:11:38 crc kubenswrapper[4667]: I0929 17:11:38.219941 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerStarted","Data":"1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d"} Sep 29 17:11:38 crc kubenswrapper[4667]: I0929 17:11:38.220263 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:11:38 crc kubenswrapper[4667]: I0929 17:11:38.238572 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podStartSLOduration=108.238561307 podStartE2EDuration="1m48.238561307s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:38.238294787 +0000 UTC m=+126.736141556" watchObservedRunningTime="2025-09-29 17:11:38.238561307 +0000 UTC m=+126.736408076" Sep 29 17:11:38 crc kubenswrapper[4667]: I0929 17:11:38.425203 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-cl5p9"] Sep 29 17:11:38 crc kubenswrapper[4667]: I0929 17:11:38.425282 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:38 crc kubenswrapper[4667]: E0929 17:11:38.425351 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:39 crc kubenswrapper[4667]: I0929 17:11:39.814982 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:39 crc kubenswrapper[4667]: E0929 17:11:39.815652 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:39 crc kubenswrapper[4667]: I0929 17:11:39.815025 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:39 crc kubenswrapper[4667]: I0929 17:11:39.815059 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:39 crc kubenswrapper[4667]: I0929 17:11:39.815002 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:39 crc kubenswrapper[4667]: E0929 17:11:39.815914 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:39 crc kubenswrapper[4667]: E0929 17:11:39.816006 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:39 crc kubenswrapper[4667]: E0929 17:11:39.816078 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:41 crc kubenswrapper[4667]: I0929 17:11:41.814927 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:41 crc kubenswrapper[4667]: E0929 17:11:41.816009 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Sep 29 17:11:41 crc kubenswrapper[4667]: I0929 17:11:41.816063 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:41 crc kubenswrapper[4667]: I0929 17:11:41.816100 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:41 crc kubenswrapper[4667]: I0929 17:11:41.816098 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:41 crc kubenswrapper[4667]: E0929 17:11:41.816176 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-cl5p9" podUID="d360e6c4-2b40-4214-bb7c-5d08038c1b62" Sep 29 17:11:41 crc kubenswrapper[4667]: E0929 17:11:41.816285 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Sep 29 17:11:41 crc kubenswrapper[4667]: E0929 17:11:41.816332 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.815108 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.815169 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.815224 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.815305 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.816616 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.816664 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.817814 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.817955 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.818197 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Sep 29 17:11:43 crc kubenswrapper[4667]: I0929 17:11:43.818245 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.399836 4667 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.420362 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5l58z"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.420756 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.421012 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j5pwf"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.421404 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.421550 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.421824 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.422008 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.422736 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.422867 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.422942 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.423188 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.424103 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.424106 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.424176 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.424415 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.424881 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.425017 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.425271 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.425392 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.425394 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.425534 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.425971 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.426097 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.426289 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.426563 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fmmvk"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.426798 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.427151 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z6xxn"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.427379 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.430730 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.441641 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.441870 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.442479 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.443583 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-qhc55"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.443941 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.444232 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.445104 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.445414 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.445581 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446010 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446044 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446063 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446174 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446219 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: W0929 17:11:45.446226 4667 reflector.go:561] object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4": failed to list *v1.Secret: secrets "machine-approver-sa-dockercfg-nl2j4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446075 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: E0929 17:11:45.446259 4667 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-nl2j4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-approver-sa-dockercfg-nl2j4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446265 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446297 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446379 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446437 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.446921 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.447008 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.447081 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.447181 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.447296 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.447454 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.447552 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.448101 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.448610 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.450681 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.450933 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.451354 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.451515 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.451553 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.451615 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.451661 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.451759 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.453219 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.453228 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.453257 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.453259 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.454471 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-b89jg"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.454579 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.454732 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.454793 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b89jg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.456445 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.456687 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.457071 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.460605 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.460622 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.460837 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.460878 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.460879 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.461287 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.461395 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.462025 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.462875 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.463854 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b5bmf"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.463944 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.464800 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.473274 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.474065 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.474276 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.474575 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.475168 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j5pwf"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.476266 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.476722 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.476994 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5l58z"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.477349 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.477648 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.478902 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f44b9"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.479506 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.479699 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2fnkw"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.480223 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.481168 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rj4qf"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.481890 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.489278 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.489700 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.489812 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490322 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490397 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490444 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490326 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490704 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490721 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490759 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490803 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490865 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490915 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490937 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490988 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.490996 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491026 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491068 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491096 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491136 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491200 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491216 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491250 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491277 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491507 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.491593 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.492572 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.492622 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.492963 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.493129 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.505053 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lwt9b"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.505803 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.506170 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.506377 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.506480 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.506541 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.507476 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.508082 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.514324 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.514724 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.514753 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.514910 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.514723 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.515410 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.515785 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.522546 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.522873 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.522984 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.523953 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.524081 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7ssxc"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.524571 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.525789 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.526122 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.526162 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vkr48"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.526601 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.526754 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.527219 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.527651 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.528003 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.528184 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.530835 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.530885 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.531589 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.532042 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.532308 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z6xxn"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.533401 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wjdtk"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.533816 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.533888 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.534216 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.536124 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.536494 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.536719 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.537049 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.543814 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-lbx5r"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.545055 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.545622 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.545994 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fmmvk"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546023 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546053 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546512 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-config\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546541 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c550aca9-a2c1-4251-8a19-e03bd9dada5e-encryption-config\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546561 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tftvg\" (UniqueName: \"kubernetes.io/projected/e0752d76-2a38-4092-ae1e-d5a9d4b395e1-kube-api-access-tftvg\") pod \"downloads-7954f5f757-b89jg\" (UID: \"e0752d76-2a38-4092-ae1e-d5a9d4b395e1\") " pod="openshift-console/downloads-7954f5f757-b89jg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546584 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546601 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/323592f8-1556-46c0-82d8-0d9fd867ab33-audit-policies\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546616 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/323592f8-1556-46c0-82d8-0d9fd867ab33-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546632 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e26e33e-dec9-47a8-9293-ebb1bfb80741-config\") pod \"kube-apiserver-operator-766d6c64bb-zmnzc\" (UID: \"5e26e33e-dec9-47a8-9293-ebb1bfb80741\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546641 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546646 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c550aca9-a2c1-4251-8a19-e03bd9dada5e-serving-cert\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546685 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-client-ca\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546702 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snvhm\" (UniqueName: \"kubernetes.io/projected/c28f53ca-7569-437f-9d3f-9eba75552b58-kube-api-access-snvhm\") pod \"openshift-apiserver-operator-796bbdcf4f-2szrg\" (UID: \"c28f53ca-7569-437f-9d3f-9eba75552b58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546731 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546746 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d2a7446-ee66-47b9-b821-fadc3fc95d32-serving-cert\") pod \"openshift-config-operator-7777fb866f-gtjfr\" (UID: \"0d2a7446-ee66-47b9-b821-fadc3fc95d32\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546760 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546774 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltbw4\" (UniqueName: \"kubernetes.io/projected/ba3449da-b041-48dd-8f39-e035467b63bf-kube-api-access-ltbw4\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546789 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq797\" (UniqueName: \"kubernetes.io/projected/0d2a7446-ee66-47b9-b821-fadc3fc95d32-kube-api-access-lq797\") pod \"openshift-config-operator-7777fb866f-gtjfr\" (UID: \"0d2a7446-ee66-47b9-b821-fadc3fc95d32\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546803 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-config\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546816 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-etcd-serving-ca\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546836 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a6113d-cddb-48e4-9855-73314af480f7-serving-cert\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546882 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546898 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-bound-sa-token\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546911 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e26e33e-dec9-47a8-9293-ebb1bfb80741-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zmnzc\" (UID: \"5e26e33e-dec9-47a8-9293-ebb1bfb80741\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546925 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs2s9\" (UniqueName: \"kubernetes.io/projected/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-kube-api-access-bs2s9\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546939 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-config\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546953 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546965 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c550aca9-a2c1-4251-8a19-e03bd9dada5e-etcd-client\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.546978 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmtwf\" (UniqueName: \"kubernetes.io/projected/c550aca9-a2c1-4251-8a19-e03bd9dada5e-kube-api-access-rmtwf\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547009 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz8pj\" (UniqueName: \"kubernetes.io/projected/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-kube-api-access-kz8pj\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547037 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/04c30c9d-8ef3-45b0-ba61-695d313226ca-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tz7dc\" (UID: \"04c30c9d-8ef3-45b0-ba61-695d313226ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547053 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ba3449da-b041-48dd-8f39-e035467b63bf-proxy-tls\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547067 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbadde10-9706-48b9-947a-f457520949d2-config\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547080 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547093 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c550aca9-a2c1-4251-8a19-e03bd9dada5e-node-pullsecrets\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547108 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn99j\" (UniqueName: \"kubernetes.io/projected/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-kube-api-access-sn99j\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547121 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/323592f8-1556-46c0-82d8-0d9fd867ab33-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547134 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-policies\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547148 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547162 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-service-ca-bundle\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547175 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c28f53ca-7569-437f-9d3f-9eba75552b58-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2szrg\" (UID: \"c28f53ca-7569-437f-9d3f-9eba75552b58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547196 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbadde10-9706-48b9-947a-f457520949d2-serving-cert\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547209 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c28f53ca-7569-437f-9d3f-9eba75552b58-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2szrg\" (UID: \"c28f53ca-7569-437f-9d3f-9eba75552b58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547221 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-serving-cert\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547235 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547249 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547261 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/323592f8-1556-46c0-82d8-0d9fd867ab33-audit-dir\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547273 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ba3449da-b041-48dd-8f39-e035467b63bf-images\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547288 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bbadde10-9706-48b9-947a-f457520949d2-etcd-ca\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547309 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bbadde10-9706-48b9-947a-f457520949d2-etcd-client\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547327 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547340 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrdrs\" (UniqueName: \"kubernetes.io/projected/80a6113d-cddb-48e4-9855-73314af480f7-kube-api-access-jrdrs\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547355 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wgjs\" (UniqueName: \"kubernetes.io/projected/8c91ba86-0e55-462b-acf3-08d7308d7df8-kube-api-access-5wgjs\") pod \"openshift-controller-manager-operator-756b6f6bc6-fflcd\" (UID: \"8c91ba86-0e55-462b-acf3-08d7308d7df8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547369 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0d2a7446-ee66-47b9-b821-fadc3fc95d32-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gtjfr\" (UID: \"0d2a7446-ee66-47b9-b821-fadc3fc95d32\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547381 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323592f8-1556-46c0-82d8-0d9fd867ab33-serving-cert\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547395 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-config\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547409 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547422 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8e68faca-8329-4201-a26d-e90968c0a8fb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-q78dk\" (UID: \"8e68faca-8329-4201-a26d-e90968c0a8fb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547448 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-trusted-ca\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547461 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh9ft\" (UniqueName: \"kubernetes.io/projected/bbadde10-9706-48b9-947a-f457520949d2-kube-api-access-wh9ft\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547482 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ba3449da-b041-48dd-8f39-e035467b63bf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547495 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c91ba86-0e55-462b-acf3-08d7308d7df8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fflcd\" (UID: \"8c91ba86-0e55-462b-acf3-08d7308d7df8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547507 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-dir\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547522 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c550aca9-a2c1-4251-8a19-e03bd9dada5e-audit-dir\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547553 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-config\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547577 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-serving-cert\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547599 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/323592f8-1556-46c0-82d8-0d9fd867ab33-etcd-client\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547618 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e26e33e-dec9-47a8-9293-ebb1bfb80741-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zmnzc\" (UID: \"5e26e33e-dec9-47a8-9293-ebb1bfb80741\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547640 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547656 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdn89\" (UniqueName: \"kubernetes.io/projected/8e68faca-8329-4201-a26d-e90968c0a8fb-kube-api-access-pdn89\") pod \"cluster-samples-operator-665b6dd947-q78dk\" (UID: \"8e68faca-8329-4201-a26d-e90968c0a8fb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547670 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04c30c9d-8ef3-45b0-ba61-695d313226ca-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tz7dc\" (UID: \"04c30c9d-8ef3-45b0-ba61-695d313226ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547685 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547694 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04c30c9d-8ef3-45b0-ba61-695d313226ca-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tz7dc\" (UID: \"04c30c9d-8ef3-45b0-ba61-695d313226ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547714 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vl6r\" (UniqueName: \"kubernetes.io/projected/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-kube-api-access-8vl6r\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547742 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbadde10-9706-48b9-947a-f457520949d2-etcd-service-ca\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547756 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-metrics-tls\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547769 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547789 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c91ba86-0e55-462b-acf3-08d7308d7df8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fflcd\" (UID: \"8c91ba86-0e55-462b-acf3-08d7308d7df8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547803 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7nbt\" (UniqueName: \"kubernetes.io/projected/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-kube-api-access-z7nbt\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547859 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-trusted-ca\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547882 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/323592f8-1556-46c0-82d8-0d9fd867ab33-encryption-config\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547899 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpg85\" (UniqueName: \"kubernetes.io/projected/323592f8-1556-46c0-82d8-0d9fd867ab33-kube-api-access-vpg85\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.547913 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-audit\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.550208 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-images\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.550245 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-image-import-ca\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.550279 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.550456 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b5bmf"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.552558 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2fnkw"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.553699 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.555059 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.556903 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.556930 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b89jg"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.558350 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-qhc55"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.558649 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.559398 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.560147 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.567169 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f44b9"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.568869 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.569346 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rj4qf"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.569622 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.571744 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.574243 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vkr48"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.575281 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qn7ls"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.576155 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.576266 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.577327 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-lbx5r"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.578080 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.580113 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7ssxc"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.580952 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.581936 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.582655 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wjdtk"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.583792 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.584295 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.585147 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qn7ls"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.586414 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.586633 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.586779 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.587648 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.588521 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.589514 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-xbcfz"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.590170 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.590365 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-tr4hr"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.590767 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tr4hr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.591250 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xbcfz"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.592060 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tr4hr"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.602853 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.622069 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.639463 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-hhstg"] Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.639992 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.642938 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650775 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/323592f8-1556-46c0-82d8-0d9fd867ab33-audit-policies\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650802 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/323592f8-1556-46c0-82d8-0d9fd867ab33-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650821 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e26e33e-dec9-47a8-9293-ebb1bfb80741-config\") pod \"kube-apiserver-operator-766d6c64bb-zmnzc\" (UID: \"5e26e33e-dec9-47a8-9293-ebb1bfb80741\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650853 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-client-ca\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650868 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c550aca9-a2c1-4251-8a19-e03bd9dada5e-serving-cert\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650889 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650905 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snvhm\" (UniqueName: \"kubernetes.io/projected/c28f53ca-7569-437f-9d3f-9eba75552b58-kube-api-access-snvhm\") pod \"openshift-apiserver-operator-796bbdcf4f-2szrg\" (UID: \"c28f53ca-7569-437f-9d3f-9eba75552b58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650923 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650940 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltbw4\" (UniqueName: \"kubernetes.io/projected/ba3449da-b041-48dd-8f39-e035467b63bf-kube-api-access-ltbw4\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650954 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d2a7446-ee66-47b9-b821-fadc3fc95d32-serving-cert\") pod \"openshift-config-operator-7777fb866f-gtjfr\" (UID: \"0d2a7446-ee66-47b9-b821-fadc3fc95d32\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650968 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-config\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650983 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq797\" (UniqueName: \"kubernetes.io/projected/0d2a7446-ee66-47b9-b821-fadc3fc95d32-kube-api-access-lq797\") pod \"openshift-config-operator-7777fb866f-gtjfr\" (UID: \"0d2a7446-ee66-47b9-b821-fadc3fc95d32\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.650996 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a6113d-cddb-48e4-9855-73314af480f7-serving-cert\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651011 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-etcd-serving-ca\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651029 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651044 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e26e33e-dec9-47a8-9293-ebb1bfb80741-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zmnzc\" (UID: \"5e26e33e-dec9-47a8-9293-ebb1bfb80741\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651061 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-bound-sa-token\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651075 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs2s9\" (UniqueName: \"kubernetes.io/projected/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-kube-api-access-bs2s9\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651090 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-config\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651104 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz8pj\" (UniqueName: \"kubernetes.io/projected/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-kube-api-access-kz8pj\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651118 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/04c30c9d-8ef3-45b0-ba61-695d313226ca-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tz7dc\" (UID: \"04c30c9d-8ef3-45b0-ba61-695d313226ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651134 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651147 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c550aca9-a2c1-4251-8a19-e03bd9dada5e-etcd-client\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651160 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmtwf\" (UniqueName: \"kubernetes.io/projected/c550aca9-a2c1-4251-8a19-e03bd9dada5e-kube-api-access-rmtwf\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651176 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbadde10-9706-48b9-947a-f457520949d2-config\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651190 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651205 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c550aca9-a2c1-4251-8a19-e03bd9dada5e-node-pullsecrets\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651221 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ba3449da-b041-48dd-8f39-e035467b63bf-proxy-tls\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651236 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn99j\" (UniqueName: \"kubernetes.io/projected/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-kube-api-access-sn99j\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651250 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651265 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-service-ca-bundle\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651278 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/323592f8-1556-46c0-82d8-0d9fd867ab33-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651294 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-policies\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651310 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbadde10-9706-48b9-947a-f457520949d2-serving-cert\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651326 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c28f53ca-7569-437f-9d3f-9eba75552b58-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2szrg\" (UID: \"c28f53ca-7569-437f-9d3f-9eba75552b58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651343 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651358 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651373 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c28f53ca-7569-437f-9d3f-9eba75552b58-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2szrg\" (UID: \"c28f53ca-7569-437f-9d3f-9eba75552b58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651390 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-serving-cert\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651404 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/323592f8-1556-46c0-82d8-0d9fd867ab33-audit-dir\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651418 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ba3449da-b041-48dd-8f39-e035467b63bf-images\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651443 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bbadde10-9706-48b9-947a-f457520949d2-etcd-ca\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651463 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bbadde10-9706-48b9-947a-f457520949d2-etcd-client\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651481 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651495 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrdrs\" (UniqueName: \"kubernetes.io/projected/80a6113d-cddb-48e4-9855-73314af480f7-kube-api-access-jrdrs\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651509 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0d2a7446-ee66-47b9-b821-fadc3fc95d32-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gtjfr\" (UID: \"0d2a7446-ee66-47b9-b821-fadc3fc95d32\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651523 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323592f8-1556-46c0-82d8-0d9fd867ab33-serving-cert\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651536 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-config\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651553 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wgjs\" (UniqueName: \"kubernetes.io/projected/8c91ba86-0e55-462b-acf3-08d7308d7df8-kube-api-access-5wgjs\") pod \"openshift-controller-manager-operator-756b6f6bc6-fflcd\" (UID: \"8c91ba86-0e55-462b-acf3-08d7308d7df8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651568 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651592 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8e68faca-8329-4201-a26d-e90968c0a8fb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-q78dk\" (UID: \"8e68faca-8329-4201-a26d-e90968c0a8fb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651610 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-trusted-ca\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651624 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh9ft\" (UniqueName: \"kubernetes.io/projected/bbadde10-9706-48b9-947a-f457520949d2-kube-api-access-wh9ft\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651640 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ba3449da-b041-48dd-8f39-e035467b63bf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651655 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c91ba86-0e55-462b-acf3-08d7308d7df8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fflcd\" (UID: \"8c91ba86-0e55-462b-acf3-08d7308d7df8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651670 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-config\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651687 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-serving-cert\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651701 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/323592f8-1556-46c0-82d8-0d9fd867ab33-etcd-client\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651715 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-dir\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651732 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c550aca9-a2c1-4251-8a19-e03bd9dada5e-audit-dir\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651748 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651765 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdn89\" (UniqueName: \"kubernetes.io/projected/8e68faca-8329-4201-a26d-e90968c0a8fb-kube-api-access-pdn89\") pod \"cluster-samples-operator-665b6dd947-q78dk\" (UID: \"8e68faca-8329-4201-a26d-e90968c0a8fb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651780 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04c30c9d-8ef3-45b0-ba61-695d313226ca-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tz7dc\" (UID: \"04c30c9d-8ef3-45b0-ba61-695d313226ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651794 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e26e33e-dec9-47a8-9293-ebb1bfb80741-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zmnzc\" (UID: \"5e26e33e-dec9-47a8-9293-ebb1bfb80741\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651809 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04c30c9d-8ef3-45b0-ba61-695d313226ca-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tz7dc\" (UID: \"04c30c9d-8ef3-45b0-ba61-695d313226ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651824 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vl6r\" (UniqueName: \"kubernetes.io/projected/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-kube-api-access-8vl6r\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651852 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651870 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c91ba86-0e55-462b-acf3-08d7308d7df8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fflcd\" (UID: \"8c91ba86-0e55-462b-acf3-08d7308d7df8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651883 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbadde10-9706-48b9-947a-f457520949d2-etcd-service-ca\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651897 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-metrics-tls\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651919 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-trusted-ca\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651935 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7nbt\" (UniqueName: \"kubernetes.io/projected/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-kube-api-access-z7nbt\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651951 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/323592f8-1556-46c0-82d8-0d9fd867ab33-encryption-config\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651965 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpg85\" (UniqueName: \"kubernetes.io/projected/323592f8-1556-46c0-82d8-0d9fd867ab33-kube-api-access-vpg85\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651981 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-audit\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.651997 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-images\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652011 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-image-import-ca\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652025 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652042 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c550aca9-a2c1-4251-8a19-e03bd9dada5e-encryption-config\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652056 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-config\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652072 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tftvg\" (UniqueName: \"kubernetes.io/projected/e0752d76-2a38-4092-ae1e-d5a9d4b395e1-kube-api-access-tftvg\") pod \"downloads-7954f5f757-b89jg\" (UID: \"e0752d76-2a38-4092-ae1e-d5a9d4b395e1\") " pod="openshift-console/downloads-7954f5f757-b89jg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652095 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652103 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/323592f8-1556-46c0-82d8-0d9fd867ab33-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652264 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652342 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-policies\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652376 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-client-ca\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652421 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-config\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652587 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/323592f8-1556-46c0-82d8-0d9fd867ab33-audit-policies\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.652927 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-config\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.653469 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-etcd-serving-ca\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.653783 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c550aca9-a2c1-4251-8a19-e03bd9dada5e-audit-dir\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.653788 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.654090 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-dir\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.654130 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-service-ca-bundle\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.653473 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c550aca9-a2c1-4251-8a19-e03bd9dada5e-node-pullsecrets\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.654282 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.654310 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/323592f8-1556-46c0-82d8-0d9fd867ab33-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.654294 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/bbadde10-9706-48b9-947a-f457520949d2-etcd-service-ca\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.654453 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/323592f8-1556-46c0-82d8-0d9fd867ab33-audit-dir\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.654674 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0d2a7446-ee66-47b9-b821-fadc3fc95d32-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gtjfr\" (UID: \"0d2a7446-ee66-47b9-b821-fadc3fc95d32\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.655078 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bbadde10-9706-48b9-947a-f457520949d2-config\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.655129 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ba3449da-b041-48dd-8f39-e035467b63bf-auth-proxy-config\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.655167 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-config\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.655525 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-image-import-ca\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.655591 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-trusted-ca\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.655667 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/bbadde10-9706-48b9-947a-f457520949d2-etcd-ca\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.655688 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c28f53ca-7569-437f-9d3f-9eba75552b58-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2szrg\" (UID: \"c28f53ca-7569-437f-9d3f-9eba75552b58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.656085 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a6113d-cddb-48e4-9855-73314af480f7-serving-cert\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.656141 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-images\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.656165 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-serving-cert\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.657017 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.657328 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.657734 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-config\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.657813 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bbadde10-9706-48b9-947a-f457520949d2-etcd-client\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.657814 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8e68faca-8329-4201-a26d-e90968c0a8fb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-q78dk\" (UID: \"8e68faca-8329-4201-a26d-e90968c0a8fb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.658007 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c550aca9-a2c1-4251-8a19-e03bd9dada5e-serving-cert\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.658567 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.658702 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.658751 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.658883 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c28f53ca-7569-437f-9d3f-9eba75552b58-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2szrg\" (UID: \"c28f53ca-7569-437f-9d3f-9eba75552b58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.659029 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d2a7446-ee66-47b9-b821-fadc3fc95d32-serving-cert\") pod \"openshift-config-operator-7777fb866f-gtjfr\" (UID: \"0d2a7446-ee66-47b9-b821-fadc3fc95d32\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.659197 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/323592f8-1556-46c0-82d8-0d9fd867ab33-encryption-config\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.659604 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.659881 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-serving-cert\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.659980 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.659987 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/323592f8-1556-46c0-82d8-0d9fd867ab33-serving-cert\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.660170 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.660233 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.660381 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.660479 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bbadde10-9706-48b9-947a-f457520949d2-serving-cert\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.660518 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c550aca9-a2c1-4251-8a19-e03bd9dada5e-etcd-client\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.660814 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/323592f8-1556-46c0-82d8-0d9fd867ab33-etcd-client\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.661467 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.662248 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.667967 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c550aca9-a2c1-4251-8a19-e03bd9dada5e-encryption-config\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.682633 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.684817 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-config\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.701757 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.706909 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c550aca9-a2c1-4251-8a19-e03bd9dada5e-audit\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.721887 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.742025 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.761829 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.766752 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c91ba86-0e55-462b-acf3-08d7308d7df8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-fflcd\" (UID: \"8c91ba86-0e55-462b-acf3-08d7308d7df8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.782382 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.784800 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c91ba86-0e55-462b-acf3-08d7308d7df8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-fflcd\" (UID: \"8c91ba86-0e55-462b-acf3-08d7308d7df8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.802064 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.822781 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.842861 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.862733 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.866688 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e26e33e-dec9-47a8-9293-ebb1bfb80741-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zmnzc\" (UID: \"5e26e33e-dec9-47a8-9293-ebb1bfb80741\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.882412 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.882857 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e26e33e-dec9-47a8-9293-ebb1bfb80741-config\") pod \"kube-apiserver-operator-766d6c64bb-zmnzc\" (UID: \"5e26e33e-dec9-47a8-9293-ebb1bfb80741\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.901919 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.922385 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.942822 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.962262 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Sep 29 17:11:45 crc kubenswrapper[4667]: I0929 17:11:45.982151 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.001825 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.022595 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.042345 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.062447 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.081980 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.101983 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.122197 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.141897 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.162537 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.182188 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.202675 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.222203 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.242100 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.246477 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04c30c9d-8ef3-45b0-ba61-695d313226ca-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tz7dc\" (UID: \"04c30c9d-8ef3-45b0-ba61-695d313226ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.263230 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.264384 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04c30c9d-8ef3-45b0-ba61-695d313226ca-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tz7dc\" (UID: \"04c30c9d-8ef3-45b0-ba61-695d313226ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.282865 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.302340 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.306606 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-metrics-tls\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.327280 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.335146 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-trusted-ca\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.342628 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.362509 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.382018 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.402920 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.406468 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ba3449da-b041-48dd-8f39-e035467b63bf-images\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.442768 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.445772 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ba3449da-b041-48dd-8f39-e035467b63bf-proxy-tls\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.462260 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.504372 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.522506 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.541283 4667 request.go:700] Waited for 1.01649559s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/secrets?fieldSelector=metadata.name%3Dmetrics-tls&limit=500&resourceVersion=0 Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.542700 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.562043 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.583060 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.602581 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.622199 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.643009 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.662519 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.682715 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.702330 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.722394 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.742640 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.762739 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.782144 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.802056 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.821992 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.842260 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.862500 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.882093 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.901965 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.922230 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.949017 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.962453 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Sep 29 17:11:46 crc kubenswrapper[4667]: I0929 17:11:46.982895 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.002152 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.022073 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.042185 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.062215 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.082187 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.101910 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.122636 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.141946 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.162263 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.182074 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.202405 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.222489 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.241931 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.263142 4667 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.281783 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.302211 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.322299 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.342491 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.362551 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.382635 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.402031 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.421998 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.441943 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.462516 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.482733 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.502145 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.533977 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz8pj\" (UniqueName: \"kubernetes.io/projected/5fb62347-e9c8-4815-a631-8fe0b5c78bd3-kube-api-access-kz8pj\") pod \"machine-api-operator-5694c8668f-j5pwf\" (UID: \"5fb62347-e9c8-4815-a631-8fe0b5c78bd3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.541983 4667 request.go:700] Waited for 1.890304555s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver-operator/serviceaccounts/kube-apiserver-operator/token Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.541995 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.553932 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5e26e33e-dec9-47a8-9293-ebb1bfb80741-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zmnzc\" (UID: \"5e26e33e-dec9-47a8-9293-ebb1bfb80741\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.572796 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-bound-sa-token\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.606994 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq797\" (UniqueName: \"kubernetes.io/projected/0d2a7446-ee66-47b9-b821-fadc3fc95d32-kube-api-access-lq797\") pod \"openshift-config-operator-7777fb866f-gtjfr\" (UID: \"0d2a7446-ee66-47b9-b821-fadc3fc95d32\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.619512 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs2s9\" (UniqueName: \"kubernetes.io/projected/d0b64aaf-d950-45bd-8fc7-4a6ba94c7469-kube-api-access-bs2s9\") pod \"authentication-operator-69f744f599-fmmvk\" (UID: \"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.626180 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.632815 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltbw4\" (UniqueName: \"kubernetes.io/projected/ba3449da-b041-48dd-8f39-e035467b63bf-kube-api-access-ltbw4\") pod \"machine-config-operator-74547568cd-tzmmc\" (UID: \"ba3449da-b041-48dd-8f39-e035467b63bf\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.653832 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn99j\" (UniqueName: \"kubernetes.io/projected/e3a95ccd-fb36-4d54-bac9-4ec54e100d43-kube-api-access-sn99j\") pod \"ingress-operator-5b745b69d9-h5s6w\" (UID: \"e3a95ccd-fb36-4d54-bac9-4ec54e100d43\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.673220 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-j5pwf"] Sep 29 17:11:47 crc kubenswrapper[4667]: W0929 17:11:47.675281 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fb62347_e9c8_4815_a631_8fe0b5c78bd3.slice/crio-66f084b9d898411a4f12b83a2d88aa846459af7c1da42c3e35dfd4198632f8d4 WatchSource:0}: Error finding container 66f084b9d898411a4f12b83a2d88aa846459af7c1da42c3e35dfd4198632f8d4: Status 404 returned error can't find the container with id 66f084b9d898411a4f12b83a2d88aa846459af7c1da42c3e35dfd4198632f8d4 Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.676000 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/04c30c9d-8ef3-45b0-ba61-695d313226ca-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tz7dc\" (UID: \"04c30c9d-8ef3-45b0-ba61-695d313226ca\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.695580 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snvhm\" (UniqueName: \"kubernetes.io/projected/c28f53ca-7569-437f-9d3f-9eba75552b58-kube-api-access-snvhm\") pod \"openshift-apiserver-operator-796bbdcf4f-2szrg\" (UID: \"c28f53ca-7569-437f-9d3f-9eba75552b58\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.713462 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7nbt\" (UniqueName: \"kubernetes.io/projected/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-kube-api-access-z7nbt\") pod \"oauth-openshift-558db77b4-z6xxn\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.735528 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmtwf\" (UniqueName: \"kubernetes.io/projected/c550aca9-a2c1-4251-8a19-e03bd9dada5e-kube-api-access-rmtwf\") pod \"apiserver-76f77b778f-rj4qf\" (UID: \"c550aca9-a2c1-4251-8a19-e03bd9dada5e\") " pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.738637 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.741504 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr"] Sep 29 17:11:47 crc kubenswrapper[4667]: W0929 17:11:47.747263 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d2a7446_ee66_47b9_b821_fadc3fc95d32.slice/crio-9ec7bbc8194cdf126bb3be404819f767744d82816cd82227070c6ad3ced5c37f WatchSource:0}: Error finding container 9ec7bbc8194cdf126bb3be404819f767744d82816cd82227070c6ad3ced5c37f: Status 404 returned error can't find the container with id 9ec7bbc8194cdf126bb3be404819f767744d82816cd82227070c6ad3ced5c37f Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.755408 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vl6r\" (UniqueName: \"kubernetes.io/projected/ee9d8867-95b1-4b66-ac9b-3b0f582c8f61-kube-api-access-8vl6r\") pod \"console-operator-58897d9998-b5bmf\" (UID: \"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61\") " pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.773252 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrdrs\" (UniqueName: \"kubernetes.io/projected/80a6113d-cddb-48e4-9855-73314af480f7-kube-api-access-jrdrs\") pod \"route-controller-manager-6576b87f9c-8qlvp\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.795041 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdn89\" (UniqueName: \"kubernetes.io/projected/8e68faca-8329-4201-a26d-e90968c0a8fb-kube-api-access-pdn89\") pod \"cluster-samples-operator-665b6dd947-q78dk\" (UID: \"8e68faca-8329-4201-a26d-e90968c0a8fb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.800561 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.807145 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.811470 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.814528 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpg85\" (UniqueName: \"kubernetes.io/projected/323592f8-1556-46c0-82d8-0d9fd867ab33-kube-api-access-vpg85\") pod \"apiserver-7bbb656c7d-pk8c9\" (UID: \"323592f8-1556-46c0-82d8-0d9fd867ab33\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.834228 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wgjs\" (UniqueName: \"kubernetes.io/projected/8c91ba86-0e55-462b-acf3-08d7308d7df8-kube-api-access-5wgjs\") pod \"openshift-controller-manager-operator-756b6f6bc6-fflcd\" (UID: \"8c91ba86-0e55-462b-acf3-08d7308d7df8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.849180 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc"] Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.850694 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.854655 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh9ft\" (UniqueName: \"kubernetes.io/projected/bbadde10-9706-48b9-947a-f457520949d2-kube-api-access-wh9ft\") pod \"etcd-operator-b45778765-2fnkw\" (UID: \"bbadde10-9706-48b9-947a-f457520949d2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.857552 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:47 crc kubenswrapper[4667]: W0929 17:11:47.864363 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e26e33e_dec9_47a8_9293_ebb1bfb80741.slice/crio-1b45be37cd2c4f64c4dcafd1e5ee51493c88f84e7053445758fbedc1927c64d5 WatchSource:0}: Error finding container 1b45be37cd2c4f64c4dcafd1e5ee51493c88f84e7053445758fbedc1927c64d5: Status 404 returned error can't find the container with id 1b45be37cd2c4f64c4dcafd1e5ee51493c88f84e7053445758fbedc1927c64d5 Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.877032 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tftvg\" (UniqueName: \"kubernetes.io/projected/e0752d76-2a38-4092-ae1e-d5a9d4b395e1-kube-api-access-tftvg\") pod \"downloads-7954f5f757-b89jg\" (UID: \"e0752d76-2a38-4092-ae1e-d5a9d4b395e1\") " pod="openshift-console/downloads-7954f5f757-b89jg" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.884067 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.892402 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.923358 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.931697 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b89jg" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.939817 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.948074 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.968072 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976174 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-trusted-ca-bundle\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976200 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-config\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976218 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrxhg\" (UniqueName: \"kubernetes.io/projected/99797fa8-02d2-42d2-85b6-a96e05a8fe59-kube-api-access-lrxhg\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976236 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55f01238-26fc-4f87-a22b-19024c896023-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2m7w5\" (UID: \"55f01238-26fc-4f87-a22b-19024c896023\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976294 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-registry-tls\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976312 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-oauth-config\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976328 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/62a04aa6-d927-4890-b405-c4dfea426f50-proxy-tls\") pod \"machine-config-controller-84d6567774-2kwqd\" (UID: \"62a04aa6-d927-4890-b405-c4dfea426f50\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976362 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eacfedb4-33c9-4744-9399-cf1848bb0353-service-ca-bundle\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976395 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdskj\" (UniqueName: \"kubernetes.io/projected/62a04aa6-d927-4890-b405-c4dfea426f50-kube-api-access-pdskj\") pod \"machine-config-controller-84d6567774-2kwqd\" (UID: \"62a04aa6-d927-4890-b405-c4dfea426f50\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976427 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d785eccf-d104-4f86-8339-884e0bbc4a52-machine-approver-tls\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976446 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsw6l\" (UniqueName: \"kubernetes.io/projected/d785eccf-d104-4f86-8339-884e0bbc4a52-kube-api-access-wsw6l\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976510 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-client-ca\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976550 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976571 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f2261944-710e-493f-9360-de4dea650ae5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976602 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/eacfedb4-33c9-4744-9399-cf1848bb0353-stats-auth\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976620 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-bound-sa-token\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976635 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eacfedb4-33c9-4744-9399-cf1848bb0353-metrics-certs\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976647 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-config\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976662 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/99797fa8-02d2-42d2-85b6-a96e05a8fe59-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976682 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/eacfedb4-33c9-4744-9399-cf1848bb0353-default-certificate\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976695 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d785eccf-d104-4f86-8339-884e0bbc4a52-auth-proxy-config\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976709 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55f01238-26fc-4f87-a22b-19024c896023-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2m7w5\" (UID: \"55f01238-26fc-4f87-a22b-19024c896023\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976732 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-trusted-ca\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976745 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6sn9\" (UniqueName: \"kubernetes.io/projected/cab548cd-ee10-421c-9648-02dd9ad58dfa-kube-api-access-q6sn9\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976758 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/99797fa8-02d2-42d2-85b6-a96e05a8fe59-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976774 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np47n\" (UniqueName: \"kubernetes.io/projected/eacfedb4-33c9-4744-9399-cf1848bb0353-kube-api-access-np47n\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976793 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-oauth-serving-cert\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976818 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-registry-certificates\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976831 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9807d262-d148-47f0-afd5-9b92aacfba38-serving-cert\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976865 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62a04aa6-d927-4890-b405-c4dfea426f50-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2kwqd\" (UID: \"62a04aa6-d927-4890-b405-c4dfea426f50\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976885 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/99797fa8-02d2-42d2-85b6-a96e05a8fe59-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976898 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55f01238-26fc-4f87-a22b-19024c896023-config\") pod \"kube-controller-manager-operator-78b949d7b-2m7w5\" (UID: \"55f01238-26fc-4f87-a22b-19024c896023\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976913 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976935 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f6dx\" (UniqueName: \"kubernetes.io/projected/3c761a0b-9b2c-4556-9e1e-060c3529ab38-kube-api-access-2f6dx\") pod \"migrator-59844c95c7-pk54t\" (UID: \"3c761a0b-9b2c-4556-9e1e-060c3529ab38\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976952 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f2261944-710e-493f-9360-de4dea650ae5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976966 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw7n4\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-kube-api-access-cw7n4\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.976979 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-service-ca\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.977000 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-serving-cert\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.977021 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d785eccf-d104-4f86-8339-884e0bbc4a52-config\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.977039 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4xhc\" (UniqueName: \"kubernetes.io/projected/9807d262-d148-47f0-afd5-9b92aacfba38-kube-api-access-w4xhc\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:47 crc kubenswrapper[4667]: E0929 17:11:47.977283 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:48.477272195 +0000 UTC m=+136.975118964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:47 crc kubenswrapper[4667]: I0929 17:11:47.985889 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.003949 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.026578 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.030896 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.078924 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079102 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/eacfedb4-33c9-4744-9399-cf1848bb0353-default-certificate\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079129 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d785eccf-d104-4f86-8339-884e0bbc4a52-auth-proxy-config\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079144 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55f01238-26fc-4f87-a22b-19024c896023-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2m7w5\" (UID: \"55f01238-26fc-4f87-a22b-19024c896023\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079166 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-trusted-ca\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079182 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6sn9\" (UniqueName: \"kubernetes.io/projected/cab548cd-ee10-421c-9648-02dd9ad58dfa-kube-api-access-q6sn9\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079197 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/99797fa8-02d2-42d2-85b6-a96e05a8fe59-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079214 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee052316-056b-4cb5-b05c-b554502091da-config\") pod \"service-ca-operator-777779d784-vb7qx\" (UID: \"ee052316-056b-4cb5-b05c-b554502091da\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079229 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np47n\" (UniqueName: \"kubernetes.io/projected/eacfedb4-33c9-4744-9399-cf1848bb0353-kube-api-access-np47n\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079246 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b20d92fc-2533-47f3-a553-7285c85ca8d0-signing-cabundle\") pod \"service-ca-9c57cc56f-lbx5r\" (UID: \"b20d92fc-2533-47f3-a553-7285c85ca8d0\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079278 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-oauth-serving-cert\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079293 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e32bf925-6e3f-4c59-bbb9-1b079a73163d-apiservice-cert\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079312 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-config-volume\") pod \"collect-profiles-29319420-lnksp\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079326 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9a4f1808-03cb-413f-bbfc-01eab48a65d1-profile-collector-cert\") pod \"catalog-operator-68c6474976-5bpkn\" (UID: \"9a4f1808-03cb-413f-bbfc-01eab48a65d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079352 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/91772b5e-ad6f-4830-8152-4e63306b448a-node-bootstrap-token\") pod \"machine-config-server-hhstg\" (UID: \"91772b5e-ad6f-4830-8152-4e63306b448a\") " pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079372 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c8035cbb-1852-41c8-8aee-820e412f9027-metrics-tls\") pod \"dns-operator-744455d44c-7ssxc\" (UID: \"c8035cbb-1852-41c8-8aee-820e412f9027\") " pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079399 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smq7v\" (UniqueName: \"kubernetes.io/projected/58bab54c-7710-4347-a85f-9b1554683c06-kube-api-access-smq7v\") pod \"olm-operator-6b444d44fb-p9kvq\" (UID: \"58bab54c-7710-4347-a85f-9b1554683c06\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079435 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-registry-certificates\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079451 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9807d262-d148-47f0-afd5-9b92aacfba38-serving-cert\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079490 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62a04aa6-d927-4890-b405-c4dfea426f50-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2kwqd\" (UID: \"62a04aa6-d927-4890-b405-c4dfea426f50\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079505 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4qm7\" (UniqueName: \"kubernetes.io/projected/3e396bd3-e718-4f53-a69b-522a601e0d4d-kube-api-access-k4qm7\") pod \"control-plane-machine-set-operator-78cbb6b69f-7x9v7\" (UID: \"3e396bd3-e718-4f53-a69b-522a601e0d4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079519 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/37d1b887-4e1d-4ea3-b247-3c6091047751-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vkr48\" (UID: \"37d1b887-4e1d-4ea3-b247-3c6091047751\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079533 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/624ad822-49b1-4cdc-851b-5727b8fdbf91-cert\") pod \"ingress-canary-tr4hr\" (UID: \"624ad822-49b1-4cdc-851b-5727b8fdbf91\") " pod="openshift-ingress-canary/ingress-canary-tr4hr" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079549 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-registration-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079563 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgtlb\" (UniqueName: \"kubernetes.io/projected/d31bb6f3-436e-47fe-b32a-be422ad9825b-kube-api-access-hgtlb\") pod \"dns-default-xbcfz\" (UID: \"d31bb6f3-436e-47fe-b32a-be422ad9825b\") " pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079580 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtgxq\" (UniqueName: \"kubernetes.io/projected/37d1b887-4e1d-4ea3-b247-3c6091047751-kube-api-access-wtgxq\") pod \"multus-admission-controller-857f4d67dd-vkr48\" (UID: \"37d1b887-4e1d-4ea3-b247-3c6091047751\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079594 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fedae4da-7b65-4159-9248-67f9f09ed3a4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rh4rn\" (UID: \"fedae4da-7b65-4159-9248-67f9f09ed3a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079606 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-plugins-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079627 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/91772b5e-ad6f-4830-8152-4e63306b448a-certs\") pod \"machine-config-server-hhstg\" (UID: \"91772b5e-ad6f-4830-8152-4e63306b448a\") " pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079660 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/99797fa8-02d2-42d2-85b6-a96e05a8fe59-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079674 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55f01238-26fc-4f87-a22b-19024c896023-config\") pod \"kube-controller-manager-operator-78b949d7b-2m7w5\" (UID: \"55f01238-26fc-4f87-a22b-19024c896023\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079688 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079713 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b20d92fc-2533-47f3-a553-7285c85ca8d0-signing-key\") pod \"service-ca-9c57cc56f-lbx5r\" (UID: \"b20d92fc-2533-47f3-a553-7285c85ca8d0\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079739 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f6dx\" (UniqueName: \"kubernetes.io/projected/3c761a0b-9b2c-4556-9e1e-060c3529ab38-kube-api-access-2f6dx\") pod \"migrator-59844c95c7-pk54t\" (UID: \"3c761a0b-9b2c-4556-9e1e-060c3529ab38\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079756 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f2261944-710e-493f-9360-de4dea650ae5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079779 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7gl9\" (UniqueName: \"kubernetes.io/projected/fedae4da-7b65-4159-9248-67f9f09ed3a4-kube-api-access-j7gl9\") pod \"kube-storage-version-migrator-operator-b67b599dd-rh4rn\" (UID: \"fedae4da-7b65-4159-9248-67f9f09ed3a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079805 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wjdtk\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079822 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/58bab54c-7710-4347-a85f-9b1554683c06-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p9kvq\" (UID: \"58bab54c-7710-4347-a85f-9b1554683c06\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079860 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw7n4\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-kube-api-access-cw7n4\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079876 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-service-ca\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079901 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/58bab54c-7710-4347-a85f-9b1554683c06-srv-cert\") pod \"olm-operator-6b444d44fb-p9kvq\" (UID: \"58bab54c-7710-4347-a85f-9b1554683c06\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079917 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-serving-cert\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.079931 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gjx8\" (UniqueName: \"kubernetes.io/projected/ee052316-056b-4cb5-b05c-b554502091da-kube-api-access-7gjx8\") pod \"service-ca-operator-777779d784-vb7qx\" (UID: \"ee052316-056b-4cb5-b05c-b554502091da\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.080134 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:48.580114494 +0000 UTC m=+137.077961263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.080610 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d785eccf-d104-4f86-8339-884e0bbc4a52-config\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.080658 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4xhc\" (UniqueName: \"kubernetes.io/projected/9807d262-d148-47f0-afd5-9b92aacfba38-kube-api-access-w4xhc\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.080736 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xb54\" (UniqueName: \"kubernetes.io/projected/624ad822-49b1-4cdc-851b-5727b8fdbf91-kube-api-access-5xb54\") pod \"ingress-canary-tr4hr\" (UID: \"624ad822-49b1-4cdc-851b-5727b8fdbf91\") " pod="openshift-ingress-canary/ingress-canary-tr4hr" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.080833 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bmcd\" (UniqueName: \"kubernetes.io/projected/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-kube-api-access-9bmcd\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.081158 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-registry-certificates\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.081162 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/99797fa8-02d2-42d2-85b6-a96e05a8fe59-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.081317 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-oauth-serving-cert\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.081410 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-service-ca\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.081514 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62a04aa6-d927-4890-b405-c4dfea426f50-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2kwqd\" (UID: \"62a04aa6-d927-4890-b405-c4dfea426f50\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.082634 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d785eccf-d104-4f86-8339-884e0bbc4a52-config\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.083245 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee052316-056b-4cb5-b05c-b554502091da-serving-cert\") pod \"service-ca-operator-777779d784-vb7qx\" (UID: \"ee052316-056b-4cb5-b05c-b554502091da\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.083274 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d785eccf-d104-4f86-8339-884e0bbc4a52-auth-proxy-config\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.083329 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-socket-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084241 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55f01238-26fc-4f87-a22b-19024c896023-config\") pod \"kube-controller-manager-operator-78b949d7b-2m7w5\" (UID: \"55f01238-26fc-4f87-a22b-19024c896023\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084248 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-trusted-ca-bundle\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084383 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-config\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084456 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtvd2\" (UniqueName: \"kubernetes.io/projected/56908f94-88ac-4298-98e5-3ba1315bd4ae-kube-api-access-vtvd2\") pod \"marketplace-operator-79b997595-wjdtk\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084484 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d31bb6f3-436e-47fe-b32a-be422ad9825b-metrics-tls\") pod \"dns-default-xbcfz\" (UID: \"d31bb6f3-436e-47fe-b32a-be422ad9825b\") " pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084525 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e32bf925-6e3f-4c59-bbb9-1b079a73163d-tmpfs\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084559 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrxhg\" (UniqueName: \"kubernetes.io/projected/99797fa8-02d2-42d2-85b6-a96e05a8fe59-kube-api-access-lrxhg\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084588 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55f01238-26fc-4f87-a22b-19024c896023-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2m7w5\" (UID: \"55f01238-26fc-4f87-a22b-19024c896023\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084766 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrlhr\" (UniqueName: \"kubernetes.io/projected/c8035cbb-1852-41c8-8aee-820e412f9027-kube-api-access-jrlhr\") pod \"dns-operator-744455d44c-7ssxc\" (UID: \"c8035cbb-1852-41c8-8aee-820e412f9027\") " pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.084915 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn7x6\" (UniqueName: \"kubernetes.io/projected/e32bf925-6e3f-4c59-bbb9-1b079a73163d-kube-api-access-sn7x6\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.085176 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-registry-tls\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.085219 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-oauth-config\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.085393 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-secret-volume\") pod \"collect-profiles-29319420-lnksp\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.085619 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/62a04aa6-d927-4890-b405-c4dfea426f50-proxy-tls\") pod \"machine-config-controller-84d6567774-2kwqd\" (UID: \"62a04aa6-d927-4890-b405-c4dfea426f50\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.085643 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d31bb6f3-436e-47fe-b32a-be422ad9825b-config-volume\") pod \"dns-default-xbcfz\" (UID: \"d31bb6f3-436e-47fe-b32a-be422ad9825b\") " pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.085933 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eacfedb4-33c9-4744-9399-cf1848bb0353-service-ca-bundle\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.086529 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-trusted-ca-bundle\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087080 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdskj\" (UniqueName: \"kubernetes.io/projected/62a04aa6-d927-4890-b405-c4dfea426f50-kube-api-access-pdskj\") pod \"machine-config-controller-84d6567774-2kwqd\" (UID: \"62a04aa6-d927-4890-b405-c4dfea426f50\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087148 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087156 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d785eccf-d104-4f86-8339-884e0bbc4a52-machine-approver-tls\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087155 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eacfedb4-33c9-4744-9399-cf1848bb0353-service-ca-bundle\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087203 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsw6l\" (UniqueName: \"kubernetes.io/projected/d785eccf-d104-4f86-8339-884e0bbc4a52-kube-api-access-wsw6l\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087229 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e396bd3-e718-4f53-a69b-522a601e0d4d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7x9v7\" (UID: \"3e396bd3-e718-4f53-a69b-522a601e0d4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087253 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcpwb\" (UniqueName: \"kubernetes.io/projected/b20d92fc-2533-47f3-a553-7285c85ca8d0-kube-api-access-mcpwb\") pod \"service-ca-9c57cc56f-lbx5r\" (UID: \"b20d92fc-2533-47f3-a553-7285c85ca8d0\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087343 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-client-ca\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087366 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf5ks\" (UniqueName: \"kubernetes.io/projected/9a4f1808-03cb-413f-bbfc-01eab48a65d1-kube-api-access-xf5ks\") pod \"catalog-operator-68c6474976-5bpkn\" (UID: \"9a4f1808-03cb-413f-bbfc-01eab48a65d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087432 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wjdtk\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087453 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47rrb\" (UniqueName: \"kubernetes.io/projected/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-kube-api-access-47rrb\") pod \"collect-profiles-29319420-lnksp\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087499 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087523 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bqjm\" (UniqueName: \"kubernetes.io/projected/5c692723-1438-4ec7-8e1a-d521d3f6e981-kube-api-access-2bqjm\") pod \"package-server-manager-789f6589d5-mfkxj\" (UID: \"5c692723-1438-4ec7-8e1a-d521d3f6e981\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087579 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-config\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087602 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e32bf925-6e3f-4c59-bbb9-1b079a73163d-webhook-cert\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087639 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f2261944-710e-493f-9360-de4dea650ae5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.087658 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-mountpoint-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.087765 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:48.587751669 +0000 UTC m=+137.085598438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.088027 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f2261944-710e-493f-9360-de4dea650ae5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.088233 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-trusted-ca\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.088326 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-client-ca\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.088443 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-csi-data-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.088493 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f2261944-710e-493f-9360-de4dea650ae5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.088687 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/eacfedb4-33c9-4744-9399-cf1848bb0353-stats-auth\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.089221 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-bound-sa-token\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.089273 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fedae4da-7b65-4159-9248-67f9f09ed3a4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rh4rn\" (UID: \"fedae4da-7b65-4159-9248-67f9f09ed3a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.089359 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eacfedb4-33c9-4744-9399-cf1848bb0353-metrics-certs\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.089453 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-config\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.089612 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c692723-1438-4ec7-8e1a-d521d3f6e981-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mfkxj\" (UID: \"5c692723-1438-4ec7-8e1a-d521d3f6e981\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.089751 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/99797fa8-02d2-42d2-85b6-a96e05a8fe59-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.089825 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9a4f1808-03cb-413f-bbfc-01eab48a65d1-srv-cert\") pod \"catalog-operator-68c6474976-5bpkn\" (UID: \"9a4f1808-03cb-413f-bbfc-01eab48a65d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.089968 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-765hp\" (UniqueName: \"kubernetes.io/projected/91772b5e-ad6f-4830-8152-4e63306b448a-kube-api-access-765hp\") pod \"machine-config-server-hhstg\" (UID: \"91772b5e-ad6f-4830-8152-4e63306b448a\") " pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.092162 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-config\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.096896 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-serving-cert\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.097187 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/eacfedb4-33c9-4744-9399-cf1848bb0353-default-certificate\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.098989 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55f01238-26fc-4f87-a22b-19024c896023-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2m7w5\" (UID: \"55f01238-26fc-4f87-a22b-19024c896023\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.099098 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9807d262-d148-47f0-afd5-9b92aacfba38-serving-cert\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.099407 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d785eccf-d104-4f86-8339-884e0bbc4a52-machine-approver-tls\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.100258 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-registry-tls\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.100931 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eacfedb4-33c9-4744-9399-cf1848bb0353-metrics-certs\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.100978 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/eacfedb4-33c9-4744-9399-cf1848bb0353-stats-auth\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.101230 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/62a04aa6-d927-4890-b405-c4dfea426f50-proxy-tls\") pod \"machine-config-controller-84d6567774-2kwqd\" (UID: \"62a04aa6-d927-4890-b405-c4dfea426f50\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.102594 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-oauth-config\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.104453 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/99797fa8-02d2-42d2-85b6-a96e05a8fe59-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.120670 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np47n\" (UniqueName: \"kubernetes.io/projected/eacfedb4-33c9-4744-9399-cf1848bb0353-kube-api-access-np47n\") pod \"router-default-5444994796-lwt9b\" (UID: \"eacfedb4-33c9-4744-9399-cf1848bb0353\") " pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.137963 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4xhc\" (UniqueName: \"kubernetes.io/projected/9807d262-d148-47f0-afd5-9b92aacfba38-kube-api-access-w4xhc\") pod \"controller-manager-879f6c89f-5l58z\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.155087 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.161630 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw7n4\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-kube-api-access-cw7n4\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.165017 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.177560 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/99797fa8-02d2-42d2-85b6-a96e05a8fe59-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190490 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190633 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-config-volume\") pod \"collect-profiles-29319420-lnksp\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190654 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9a4f1808-03cb-413f-bbfc-01eab48a65d1-profile-collector-cert\") pod \"catalog-operator-68c6474976-5bpkn\" (UID: \"9a4f1808-03cb-413f-bbfc-01eab48a65d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190674 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/91772b5e-ad6f-4830-8152-4e63306b448a-node-bootstrap-token\") pod \"machine-config-server-hhstg\" (UID: \"91772b5e-ad6f-4830-8152-4e63306b448a\") " pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190689 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c8035cbb-1852-41c8-8aee-820e412f9027-metrics-tls\") pod \"dns-operator-744455d44c-7ssxc\" (UID: \"c8035cbb-1852-41c8-8aee-820e412f9027\") " pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190705 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smq7v\" (UniqueName: \"kubernetes.io/projected/58bab54c-7710-4347-a85f-9b1554683c06-kube-api-access-smq7v\") pod \"olm-operator-6b444d44fb-p9kvq\" (UID: \"58bab54c-7710-4347-a85f-9b1554683c06\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190731 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4qm7\" (UniqueName: \"kubernetes.io/projected/3e396bd3-e718-4f53-a69b-522a601e0d4d-kube-api-access-k4qm7\") pod \"control-plane-machine-set-operator-78cbb6b69f-7x9v7\" (UID: \"3e396bd3-e718-4f53-a69b-522a601e0d4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190746 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/37d1b887-4e1d-4ea3-b247-3c6091047751-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vkr48\" (UID: \"37d1b887-4e1d-4ea3-b247-3c6091047751\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190761 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/624ad822-49b1-4cdc-851b-5727b8fdbf91-cert\") pod \"ingress-canary-tr4hr\" (UID: \"624ad822-49b1-4cdc-851b-5727b8fdbf91\") " pod="openshift-ingress-canary/ingress-canary-tr4hr" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190775 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-registration-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190789 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgtlb\" (UniqueName: \"kubernetes.io/projected/d31bb6f3-436e-47fe-b32a-be422ad9825b-kube-api-access-hgtlb\") pod \"dns-default-xbcfz\" (UID: \"d31bb6f3-436e-47fe-b32a-be422ad9825b\") " pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190803 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtgxq\" (UniqueName: \"kubernetes.io/projected/37d1b887-4e1d-4ea3-b247-3c6091047751-kube-api-access-wtgxq\") pod \"multus-admission-controller-857f4d67dd-vkr48\" (UID: \"37d1b887-4e1d-4ea3-b247-3c6091047751\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190816 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fedae4da-7b65-4159-9248-67f9f09ed3a4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rh4rn\" (UID: \"fedae4da-7b65-4159-9248-67f9f09ed3a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190829 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-plugins-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190859 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/91772b5e-ad6f-4830-8152-4e63306b448a-certs\") pod \"machine-config-server-hhstg\" (UID: \"91772b5e-ad6f-4830-8152-4e63306b448a\") " pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190880 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b20d92fc-2533-47f3-a553-7285c85ca8d0-signing-key\") pod \"service-ca-9c57cc56f-lbx5r\" (UID: \"b20d92fc-2533-47f3-a553-7285c85ca8d0\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190901 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7gl9\" (UniqueName: \"kubernetes.io/projected/fedae4da-7b65-4159-9248-67f9f09ed3a4-kube-api-access-j7gl9\") pod \"kube-storage-version-migrator-operator-b67b599dd-rh4rn\" (UID: \"fedae4da-7b65-4159-9248-67f9f09ed3a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190916 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wjdtk\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190931 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/58bab54c-7710-4347-a85f-9b1554683c06-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p9kvq\" (UID: \"58bab54c-7710-4347-a85f-9b1554683c06\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190950 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/58bab54c-7710-4347-a85f-9b1554683c06-srv-cert\") pod \"olm-operator-6b444d44fb-p9kvq\" (UID: \"58bab54c-7710-4347-a85f-9b1554683c06\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190969 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gjx8\" (UniqueName: \"kubernetes.io/projected/ee052316-056b-4cb5-b05c-b554502091da-kube-api-access-7gjx8\") pod \"service-ca-operator-777779d784-vb7qx\" (UID: \"ee052316-056b-4cb5-b05c-b554502091da\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.190996 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xb54\" (UniqueName: \"kubernetes.io/projected/624ad822-49b1-4cdc-851b-5727b8fdbf91-kube-api-access-5xb54\") pod \"ingress-canary-tr4hr\" (UID: \"624ad822-49b1-4cdc-851b-5727b8fdbf91\") " pod="openshift-ingress-canary/ingress-canary-tr4hr" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191011 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bmcd\" (UniqueName: \"kubernetes.io/projected/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-kube-api-access-9bmcd\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191029 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee052316-056b-4cb5-b05c-b554502091da-serving-cert\") pod \"service-ca-operator-777779d784-vb7qx\" (UID: \"ee052316-056b-4cb5-b05c-b554502091da\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191045 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-socket-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191064 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtvd2\" (UniqueName: \"kubernetes.io/projected/56908f94-88ac-4298-98e5-3ba1315bd4ae-kube-api-access-vtvd2\") pod \"marketplace-operator-79b997595-wjdtk\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191078 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d31bb6f3-436e-47fe-b32a-be422ad9825b-metrics-tls\") pod \"dns-default-xbcfz\" (UID: \"d31bb6f3-436e-47fe-b32a-be422ad9825b\") " pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191098 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e32bf925-6e3f-4c59-bbb9-1b079a73163d-tmpfs\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191131 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrlhr\" (UniqueName: \"kubernetes.io/projected/c8035cbb-1852-41c8-8aee-820e412f9027-kube-api-access-jrlhr\") pod \"dns-operator-744455d44c-7ssxc\" (UID: \"c8035cbb-1852-41c8-8aee-820e412f9027\") " pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191145 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn7x6\" (UniqueName: \"kubernetes.io/projected/e32bf925-6e3f-4c59-bbb9-1b079a73163d-kube-api-access-sn7x6\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191164 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-secret-volume\") pod \"collect-profiles-29319420-lnksp\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191179 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d31bb6f3-436e-47fe-b32a-be422ad9825b-config-volume\") pod \"dns-default-xbcfz\" (UID: \"d31bb6f3-436e-47fe-b32a-be422ad9825b\") " pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191208 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e396bd3-e718-4f53-a69b-522a601e0d4d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7x9v7\" (UID: \"3e396bd3-e718-4f53-a69b-522a601e0d4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191227 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcpwb\" (UniqueName: \"kubernetes.io/projected/b20d92fc-2533-47f3-a553-7285c85ca8d0-kube-api-access-mcpwb\") pod \"service-ca-9c57cc56f-lbx5r\" (UID: \"b20d92fc-2533-47f3-a553-7285c85ca8d0\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191244 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf5ks\" (UniqueName: \"kubernetes.io/projected/9a4f1808-03cb-413f-bbfc-01eab48a65d1-kube-api-access-xf5ks\") pod \"catalog-operator-68c6474976-5bpkn\" (UID: \"9a4f1808-03cb-413f-bbfc-01eab48a65d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191258 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wjdtk\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191276 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47rrb\" (UniqueName: \"kubernetes.io/projected/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-kube-api-access-47rrb\") pod \"collect-profiles-29319420-lnksp\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191826 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bqjm\" (UniqueName: \"kubernetes.io/projected/5c692723-1438-4ec7-8e1a-d521d3f6e981-kube-api-access-2bqjm\") pod \"package-server-manager-789f6589d5-mfkxj\" (UID: \"5c692723-1438-4ec7-8e1a-d521d3f6e981\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191867 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e32bf925-6e3f-4c59-bbb9-1b079a73163d-webhook-cert\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191888 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-mountpoint-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191906 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-csi-data-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191941 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fedae4da-7b65-4159-9248-67f9f09ed3a4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rh4rn\" (UID: \"fedae4da-7b65-4159-9248-67f9f09ed3a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191958 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c692723-1438-4ec7-8e1a-d521d3f6e981-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mfkxj\" (UID: \"5c692723-1438-4ec7-8e1a-d521d3f6e981\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.191988 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9a4f1808-03cb-413f-bbfc-01eab48a65d1-srv-cert\") pod \"catalog-operator-68c6474976-5bpkn\" (UID: \"9a4f1808-03cb-413f-bbfc-01eab48a65d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.192006 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-765hp\" (UniqueName: \"kubernetes.io/projected/91772b5e-ad6f-4830-8152-4e63306b448a-kube-api-access-765hp\") pod \"machine-config-server-hhstg\" (UID: \"91772b5e-ad6f-4830-8152-4e63306b448a\") " pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.192043 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee052316-056b-4cb5-b05c-b554502091da-config\") pod \"service-ca-operator-777779d784-vb7qx\" (UID: \"ee052316-056b-4cb5-b05c-b554502091da\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.192064 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b20d92fc-2533-47f3-a553-7285c85ca8d0-signing-cabundle\") pod \"service-ca-9c57cc56f-lbx5r\" (UID: \"b20d92fc-2533-47f3-a553-7285c85ca8d0\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.192080 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e32bf925-6e3f-4c59-bbb9-1b079a73163d-apiservice-cert\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.193034 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wjdtk\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.193108 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:48.693093997 +0000 UTC m=+137.190940766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.193365 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d31bb6f3-436e-47fe-b32a-be422ad9825b-config-volume\") pod \"dns-default-xbcfz\" (UID: \"d31bb6f3-436e-47fe-b32a-be422ad9825b\") " pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.194606 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e32bf925-6e3f-4c59-bbb9-1b079a73163d-apiservice-cert\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.195579 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/b20d92fc-2533-47f3-a553-7285c85ca8d0-signing-cabundle\") pod \"service-ca-9c57cc56f-lbx5r\" (UID: \"b20d92fc-2533-47f3-a553-7285c85ca8d0\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.196481 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee052316-056b-4cb5-b05c-b554502091da-config\") pod \"service-ca-operator-777779d784-vb7qx\" (UID: \"ee052316-056b-4cb5-b05c-b554502091da\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.196554 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-mountpoint-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.196694 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6sn9\" (UniqueName: \"kubernetes.io/projected/cab548cd-ee10-421c-9648-02dd9ad58dfa-kube-api-access-q6sn9\") pod \"console-f9d7485db-qhc55\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.196707 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-csi-data-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.196815 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-registration-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.196833 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-socket-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.196944 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-config-volume\") pod \"collect-profiles-29319420-lnksp\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.197122 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e32bf925-6e3f-4c59-bbb9-1b079a73163d-tmpfs\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.197626 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-plugins-dir\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.197890 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fedae4da-7b65-4159-9248-67f9f09ed3a4-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rh4rn\" (UID: \"fedae4da-7b65-4159-9248-67f9f09ed3a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.199752 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/58bab54c-7710-4347-a85f-9b1554683c06-srv-cert\") pod \"olm-operator-6b444d44fb-p9kvq\" (UID: \"58bab54c-7710-4347-a85f-9b1554683c06\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.200074 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fedae4da-7b65-4159-9248-67f9f09ed3a4-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rh4rn\" (UID: \"fedae4da-7b65-4159-9248-67f9f09ed3a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.200329 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9a4f1808-03cb-413f-bbfc-01eab48a65d1-srv-cert\") pod \"catalog-operator-68c6474976-5bpkn\" (UID: \"9a4f1808-03cb-413f-bbfc-01eab48a65d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.203372 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9a4f1808-03cb-413f-bbfc-01eab48a65d1-profile-collector-cert\") pod \"catalog-operator-68c6474976-5bpkn\" (UID: \"9a4f1808-03cb-413f-bbfc-01eab48a65d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.205253 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/624ad822-49b1-4cdc-851b-5727b8fdbf91-cert\") pod \"ingress-canary-tr4hr\" (UID: \"624ad822-49b1-4cdc-851b-5727b8fdbf91\") " pod="openshift-ingress-canary/ingress-canary-tr4hr" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.206686 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/b20d92fc-2533-47f3-a553-7285c85ca8d0-signing-key\") pod \"service-ca-9c57cc56f-lbx5r\" (UID: \"b20d92fc-2533-47f3-a553-7285c85ca8d0\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.207084 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/91772b5e-ad6f-4830-8152-4e63306b448a-node-bootstrap-token\") pod \"machine-config-server-hhstg\" (UID: \"91772b5e-ad6f-4830-8152-4e63306b448a\") " pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.207757 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/37d1b887-4e1d-4ea3-b247-3c6091047751-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vkr48\" (UID: \"37d1b887-4e1d-4ea3-b247-3c6091047751\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.209645 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d31bb6f3-436e-47fe-b32a-be422ad9825b-metrics-tls\") pod \"dns-default-xbcfz\" (UID: \"d31bb6f3-436e-47fe-b32a-be422ad9825b\") " pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.209789 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/5c692723-1438-4ec7-8e1a-d521d3f6e981-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mfkxj\" (UID: \"5c692723-1438-4ec7-8e1a-d521d3f6e981\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.209866 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c8035cbb-1852-41c8-8aee-820e412f9027-metrics-tls\") pod \"dns-operator-744455d44c-7ssxc\" (UID: \"c8035cbb-1852-41c8-8aee-820e412f9027\") " pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.209924 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/58bab54c-7710-4347-a85f-9b1554683c06-profile-collector-cert\") pod \"olm-operator-6b444d44fb-p9kvq\" (UID: \"58bab54c-7710-4347-a85f-9b1554683c06\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.210098 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e396bd3-e718-4f53-a69b-522a601e0d4d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-7x9v7\" (UID: \"3e396bd3-e718-4f53-a69b-522a601e0d4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.210325 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ee052316-056b-4cb5-b05c-b554502091da-serving-cert\") pod \"service-ca-operator-777779d784-vb7qx\" (UID: \"ee052316-056b-4cb5-b05c-b554502091da\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.210526 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/91772b5e-ad6f-4830-8152-4e63306b448a-certs\") pod \"machine-config-server-hhstg\" (UID: \"91772b5e-ad6f-4830-8152-4e63306b448a\") " pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.211166 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-secret-volume\") pod \"collect-profiles-29319420-lnksp\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.214224 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e32bf925-6e3f-4c59-bbb9-1b079a73163d-webhook-cert\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.214466 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wjdtk\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.217882 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.218398 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f6dx\" (UniqueName: \"kubernetes.io/projected/3c761a0b-9b2c-4556-9e1e-060c3529ab38-kube-api-access-2f6dx\") pod \"migrator-59844c95c7-pk54t\" (UID: \"3c761a0b-9b2c-4556-9e1e-060c3529ab38\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.235692 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55f01238-26fc-4f87-a22b-19024c896023-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2m7w5\" (UID: \"55f01238-26fc-4f87-a22b-19024c896023\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.247487 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" event={"ID":"04c30c9d-8ef3-45b0-ba61-695d313226ca","Type":"ContainerStarted","Data":"4f069e271e51e5d468e42bec7c6cb780ba3b42503abf3d4fc4f462ad32e57661"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.253570 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" event={"ID":"e3a95ccd-fb36-4d54-bac9-4ec54e100d43","Type":"ContainerStarted","Data":"23a78360d1c615eb0d47f8c3589d96c69975b615fe98effa77d204048061f3a0"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.257101 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrxhg\" (UniqueName: \"kubernetes.io/projected/99797fa8-02d2-42d2-85b6-a96e05a8fe59-kube-api-access-lrxhg\") pod \"cluster-image-registry-operator-dc59b4c8b-2npg8\" (UID: \"99797fa8-02d2-42d2-85b6-a96e05a8fe59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.258700 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" event={"ID":"ba3449da-b041-48dd-8f39-e035467b63bf","Type":"ContainerStarted","Data":"d59d6ee5113047e2933ca4dea74b40c54c436239ecf1c03072f55f4246376c83"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.258730 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" event={"ID":"ba3449da-b041-48dd-8f39-e035467b63bf","Type":"ContainerStarted","Data":"9b20913b975123c7891da43f03184d23ca79ab0e13130051ea29b6688225756a"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.261898 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" event={"ID":"5e26e33e-dec9-47a8-9293-ebb1bfb80741","Type":"ContainerStarted","Data":"fefcc563a3421171d0eaa646cd36aa6ab29c6cd445c6f3e645fcf88a0ea17ed9"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.261924 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" event={"ID":"5e26e33e-dec9-47a8-9293-ebb1bfb80741","Type":"ContainerStarted","Data":"1b45be37cd2c4f64c4dcafd1e5ee51493c88f84e7053445758fbedc1927c64d5"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.270012 4667 generic.go:334] "Generic (PLEG): container finished" podID="0d2a7446-ee66-47b9-b821-fadc3fc95d32" containerID="4edf081cecf9a6b3d8ad501eaedacf043d0c7bcb73fa09b220678ea8a71b1c62" exitCode=0 Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.270068 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" event={"ID":"0d2a7446-ee66-47b9-b821-fadc3fc95d32","Type":"ContainerDied","Data":"4edf081cecf9a6b3d8ad501eaedacf043d0c7bcb73fa09b220678ea8a71b1c62"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.270094 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" event={"ID":"0d2a7446-ee66-47b9-b821-fadc3fc95d32","Type":"ContainerStarted","Data":"9ec7bbc8194cdf126bb3be404819f767744d82816cd82227070c6ad3ced5c37f"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.275445 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" event={"ID":"5fb62347-e9c8-4815-a631-8fe0b5c78bd3","Type":"ContainerStarted","Data":"9136d8b28faa724c4b9cd4513e85aeca026812ce1f4fb85f9914d92d85012574"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.275476 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.275491 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" event={"ID":"5fb62347-e9c8-4815-a631-8fe0b5c78bd3","Type":"ContainerStarted","Data":"f8a226913d7f51d195015dd5ab187b284376a01534a24a777f5a5a315da782e3"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.275500 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" event={"ID":"5fb62347-e9c8-4815-a631-8fe0b5c78bd3","Type":"ContainerStarted","Data":"66f084b9d898411a4f12b83a2d88aa846459af7c1da42c3e35dfd4198632f8d4"} Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.277913 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdskj\" (UniqueName: \"kubernetes.io/projected/62a04aa6-d927-4890-b405-c4dfea426f50-kube-api-access-pdskj\") pod \"machine-config-controller-84d6567774-2kwqd\" (UID: \"62a04aa6-d927-4890-b405-c4dfea426f50\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.293782 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.296572 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:48.796556924 +0000 UTC m=+137.294403694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.302244 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2fnkw"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.302522 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsw6l\" (UniqueName: \"kubernetes.io/projected/d785eccf-d104-4f86-8339-884e0bbc4a52-kube-api-access-wsw6l\") pod \"machine-approver-56656f9798-c2zxb\" (UID: \"d785eccf-d104-4f86-8339-884e0bbc4a52\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.317453 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-bound-sa-token\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.335477 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.347116 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.351780 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.352575 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fmmvk"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.353640 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z6xxn"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.354108 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47rrb\" (UniqueName: \"kubernetes.io/projected/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-kube-api-access-47rrb\") pod \"collect-profiles-29319420-lnksp\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.379264 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xb54\" (UniqueName: \"kubernetes.io/projected/624ad822-49b1-4cdc-851b-5727b8fdbf91-kube-api-access-5xb54\") pod \"ingress-canary-tr4hr\" (UID: \"624ad822-49b1-4cdc-851b-5727b8fdbf91\") " pod="openshift-ingress-canary/ingress-canary-tr4hr" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.388030 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.393973 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.398655 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-qhc55"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.399083 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.399300 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf5ks\" (UniqueName: \"kubernetes.io/projected/9a4f1808-03cb-413f-bbfc-01eab48a65d1-kube-api-access-xf5ks\") pod \"catalog-operator-68c6474976-5bpkn\" (UID: \"9a4f1808-03cb-413f-bbfc-01eab48a65d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.399879 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:48.89986523 +0000 UTC m=+137.397712000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.419102 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcpwb\" (UniqueName: \"kubernetes.io/projected/b20d92fc-2533-47f3-a553-7285c85ca8d0-kube-api-access-mcpwb\") pod \"service-ca-9c57cc56f-lbx5r\" (UID: \"b20d92fc-2533-47f3-a553-7285c85ca8d0\") " pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.434698 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.443531 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-765hp\" (UniqueName: \"kubernetes.io/projected/91772b5e-ad6f-4830-8152-4e63306b448a-kube-api-access-765hp\") pod \"machine-config-server-hhstg\" (UID: \"91772b5e-ad6f-4830-8152-4e63306b448a\") " pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.449153 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-b5bmf"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.455394 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b89jg"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.456463 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.460706 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.461859 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.466957 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gjx8\" (UniqueName: \"kubernetes.io/projected/ee052316-056b-4cb5-b05c-b554502091da-kube-api-access-7gjx8\") pod \"service-ca-operator-777779d784-vb7qx\" (UID: \"ee052316-056b-4cb5-b05c-b554502091da\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.475742 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.476212 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bqjm\" (UniqueName: \"kubernetes.io/projected/5c692723-1438-4ec7-8e1a-d521d3f6e981-kube-api-access-2bqjm\") pod \"package-server-manager-789f6589d5-mfkxj\" (UID: \"5c692723-1438-4ec7-8e1a-d521d3f6e981\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.479705 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.485080 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.495363 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-tr4hr" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.496526 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hhstg" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.498371 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bmcd\" (UniqueName: \"kubernetes.io/projected/8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd-kube-api-access-9bmcd\") pod \"csi-hostpathplugin-qn7ls\" (UID: \"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd\") " pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.500855 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.501185 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.001172919 +0000 UTC m=+137.499019688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.519320 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smq7v\" (UniqueName: \"kubernetes.io/projected/58bab54c-7710-4347-a85f-9b1554683c06-kube-api-access-smq7v\") pod \"olm-operator-6b444d44fb-p9kvq\" (UID: \"58bab54c-7710-4347-a85f-9b1554683c06\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.549509 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rj4qf"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.554025 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgtlb\" (UniqueName: \"kubernetes.io/projected/d31bb6f3-436e-47fe-b32a-be422ad9825b-kube-api-access-hgtlb\") pod \"dns-default-xbcfz\" (UID: \"d31bb6f3-436e-47fe-b32a-be422ad9825b\") " pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.556191 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.556198 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtgxq\" (UniqueName: \"kubernetes.io/projected/37d1b887-4e1d-4ea3-b247-3c6091047751-kube-api-access-wtgxq\") pod \"multus-admission-controller-857f4d67dd-vkr48\" (UID: \"37d1b887-4e1d-4ea3-b247-3c6091047751\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.559676 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.580273 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrlhr\" (UniqueName: \"kubernetes.io/projected/c8035cbb-1852-41c8-8aee-820e412f9027-kube-api-access-jrlhr\") pod \"dns-operator-744455d44c-7ssxc\" (UID: \"c8035cbb-1852-41c8-8aee-820e412f9027\") " pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.599006 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn7x6\" (UniqueName: \"kubernetes.io/projected/e32bf925-6e3f-4c59-bbb9-1b079a73163d-kube-api-access-sn7x6\") pod \"packageserver-d55dfcdfc-dg5gl\" (UID: \"e32bf925-6e3f-4c59-bbb9-1b079a73163d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.600497 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.604475 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.604988 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.104974044 +0000 UTC m=+137.602820813 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.619634 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.622666 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtvd2\" (UniqueName: \"kubernetes.io/projected/56908f94-88ac-4298-98e5-3ba1315bd4ae-kube-api-access-vtvd2\") pod \"marketplace-operator-79b997595-wjdtk\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.636446 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.640319 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4qm7\" (UniqueName: \"kubernetes.io/projected/3e396bd3-e718-4f53-a69b-522a601e0d4d-kube-api-access-k4qm7\") pod \"control-plane-machine-set-operator-78cbb6b69f-7x9v7\" (UID: \"3e396bd3-e718-4f53-a69b-522a601e0d4d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.658239 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7gl9\" (UniqueName: \"kubernetes.io/projected/fedae4da-7b65-4159-9248-67f9f09ed3a4-kube-api-access-j7gl9\") pod \"kube-storage-version-migrator-operator-b67b599dd-rh4rn\" (UID: \"fedae4da-7b65-4159-9248-67f9f09ed3a4\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.707699 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.708069 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.208055402 +0000 UTC m=+137.705902172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.712425 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd"] Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.717341 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.740804 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.741489 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.741725 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.742499 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.743549 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.747747 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.759330 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.765138 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.792565 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.808581 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.808684 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.308664515 +0000 UTC m=+137.806511284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.808826 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.809070 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.309062855 +0000 UTC m=+137.806909624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.834555 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.910427 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.910653 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.410635834 +0000 UTC m=+137.908482602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.910719 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:48 crc kubenswrapper[4667]: I0929 17:11:48.910764 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zmnzc" podStartSLOduration=118.910749728 podStartE2EDuration="1m58.910749728s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:48.873404416 +0000 UTC m=+137.371251185" watchObservedRunningTime="2025-09-29 17:11:48.910749728 +0000 UTC m=+137.408596497" Sep 29 17:11:48 crc kubenswrapper[4667]: E0929 17:11:48.911046 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.411038872 +0000 UTC m=+137.908885641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.011423 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.012041 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.51202278 +0000 UTC m=+138.009869550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.114250 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.114647 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.614632782 +0000 UTC m=+138.112479551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.125960 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5l58z"] Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.134509 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-lbx5r"] Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.173254 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn"] Sep 29 17:11:49 crc kubenswrapper[4667]: W0929 17:11:49.208154 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9807d262_d148_47f0_afd5_9b92aacfba38.slice/crio-ef7e7eb075584f6603d9641ce5ce3feac4a4fa4781615bab0f7dce26bc3ecc5a WatchSource:0}: Error finding container ef7e7eb075584f6603d9641ce5ce3feac4a4fa4781615bab0f7dce26bc3ecc5a: Status 404 returned error can't find the container with id ef7e7eb075584f6603d9641ce5ce3feac4a4fa4781615bab0f7dce26bc3ecc5a Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.215098 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.215250 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.715231835 +0000 UTC m=+138.213078604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.215361 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.215619 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.715612132 +0000 UTC m=+138.213458901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.282820 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" event={"ID":"c550aca9-a2c1-4251-8a19-e03bd9dada5e","Type":"ContainerStarted","Data":"2ea4e8ff3a059123ba5320c88622209e2311076f3abea8d8f855842ca89b05e7"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.284358 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" event={"ID":"8e68faca-8329-4201-a26d-e90968c0a8fb","Type":"ContainerStarted","Data":"6bc22f45096e420f0defde76567ed9bd2a0103009f1f7119bc2de15f8d36c031"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.285879 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t" event={"ID":"3c761a0b-9b2c-4556-9e1e-060c3529ab38","Type":"ContainerStarted","Data":"14559386e32286e8cf5d95c42e34d9ab362beb12671e2d0ba79d9730d8540b84"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.301175 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" event={"ID":"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8","Type":"ContainerStarted","Data":"ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.301422 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.301434 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" event={"ID":"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8","Type":"ContainerStarted","Data":"77ea1044b8b4ea03f52c14faa214e849993a8ab5b8dc0d82b029eaf66f3f8b99"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.302732 4667 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-z6xxn container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.302759 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" podUID="f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.304554 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" event={"ID":"e3a95ccd-fb36-4d54-bac9-4ec54e100d43","Type":"ContainerStarted","Data":"132a61b3be12b66e8e3f83339ce4c5dcf08a7b5b237662d17406c3ddba6435a1"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.304575 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" event={"ID":"e3a95ccd-fb36-4d54-bac9-4ec54e100d43","Type":"ContainerStarted","Data":"9c76198f82d4bf11b674880d19e629418588dea36c49f0c0417a13c5130b4715"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.308045 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" event={"ID":"62a04aa6-d927-4890-b405-c4dfea426f50","Type":"ContainerStarted","Data":"da76ab33d98f58e7b39b71e44b29df2517212ef3050375ab745c925322ca53af"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.321772 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.322162 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.82208539 +0000 UTC m=+138.319932159 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.376565 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" event={"ID":"55f01238-26fc-4f87-a22b-19024c896023","Type":"ContainerStarted","Data":"d4259d33af938f0e6fba2b052e89d3d2d0c7716ff20219b471630061c930232a"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.414306 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" event={"ID":"ba3449da-b041-48dd-8f39-e035467b63bf","Type":"ContainerStarted","Data":"066391f84babde0a6a4f8b81ea54de871aa06c0fcd29919245d907b2e3aea41e"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.423334 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.423679 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:49.923666212 +0000 UTC m=+138.421512981 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.431028 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-j5pwf" podStartSLOduration=118.431014443 podStartE2EDuration="1m58.431014443s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:49.401026008 +0000 UTC m=+137.898872777" watchObservedRunningTime="2025-09-29 17:11:49.431014443 +0000 UTC m=+137.928861212" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.431865 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b89jg" event={"ID":"e0752d76-2a38-4092-ae1e-d5a9d4b395e1","Type":"ContainerStarted","Data":"99f0bd3a3279cb2dca257658f05ee1a84c83af61cc7aab67d5d5ebfacf56634a"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.432590 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b89jg" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.443446 4667 patch_prober.go:28] interesting pod/downloads-7954f5f757-b89jg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.443485 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b89jg" podUID="e0752d76-2a38-4092-ae1e-d5a9d4b395e1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.445538 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" event={"ID":"9807d262-d148-47f0-afd5-9b92aacfba38","Type":"ContainerStarted","Data":"ef7e7eb075584f6603d9641ce5ce3feac4a4fa4781615bab0f7dce26bc3ecc5a"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.449031 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" event={"ID":"80a6113d-cddb-48e4-9855-73314af480f7","Type":"ContainerStarted","Data":"4d3a7da769e75c5acbff30646b60d96418e90162cd4fe8acfa20bbaba188d5b9"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.449071 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" event={"ID":"80a6113d-cddb-48e4-9855-73314af480f7","Type":"ContainerStarted","Data":"555d15c8e84cbbecc8f3f9d243d6dba46699d1072b2f61a39a4ac6bf8b47e60d"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.450558 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.458151 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" event={"ID":"04c30c9d-8ef3-45b0-ba61-695d313226ca","Type":"ContainerStarted","Data":"76b5add3e891fc4963bbee1b75ad42a05c20f61c460fe0647b1157ed3d113fad"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.459530 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.461159 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" event={"ID":"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469","Type":"ContainerStarted","Data":"d957d880687fe2ec712c58102ba5f70cea26e79f05131218fefbf4d5cdacaea3"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.461261 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" event={"ID":"d0b64aaf-d950-45bd-8fc7-4a6ba94c7469","Type":"ContainerStarted","Data":"259b4489c64dde965f169974ee6e1904cd3f64f14e5e06391e8d1073b9629324"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.465180 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" event={"ID":"0d2a7446-ee66-47b9-b821-fadc3fc95d32","Type":"ContainerStarted","Data":"9032788d0ee51d547fc69fe578bf42b7ea867e3fe916f3de212124c5aacc992a"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.465303 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.468372 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hhstg" event={"ID":"91772b5e-ad6f-4830-8152-4e63306b448a","Type":"ContainerStarted","Data":"92cbbbba8fd7bfbe63f88090cc6cd53bb639eb04aab5c3447d80dc4c249909a3"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.494577 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" event={"ID":"bbadde10-9706-48b9-947a-f457520949d2","Type":"ContainerStarted","Data":"e4a5c16512166d9640d0dab4d98f183d75cb119c504cd7669058b0010736a8cd"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.494604 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" event={"ID":"bbadde10-9706-48b9-947a-f457520949d2","Type":"ContainerStarted","Data":"faf5ada0a3819819edc8a8b49cbbadca9a0ced6b929554ae47e145ca391f184d"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.507089 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp"] Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.508898 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" event={"ID":"c28f53ca-7569-437f-9d3f-9eba75552b58","Type":"ContainerStarted","Data":"029a4d4fcd7891f7f3c61b68eb14a3e61ebde83369f91e71c4bed186133fdd6e"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.508939 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" event={"ID":"c28f53ca-7569-437f-9d3f-9eba75552b58","Type":"ContainerStarted","Data":"36e6efadb2974d5f1869c77c987821a30d11862e56269c219a95e72d628488e4"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.530052 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.530767 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.030729972 +0000 UTC m=+138.528576741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.549172 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lwt9b" event={"ID":"eacfedb4-33c9-4744-9399-cf1848bb0353","Type":"ContainerStarted","Data":"816dfb18fc06c580d5d32f7b710cd8cb56367ff588283daada01db948a6ec167"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.549216 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lwt9b" event={"ID":"eacfedb4-33c9-4744-9399-cf1848bb0353","Type":"ContainerStarted","Data":"d295adecf455f72ca253f0117fdd566d0cc9d8f035e3b7f1cf3e6c4aaa2924ac"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.555567 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" event={"ID":"8c91ba86-0e55-462b-acf3-08d7308d7df8","Type":"ContainerStarted","Data":"7398bacf083423a6d04128b5bd22e9801f2e96c1e3e831ba33442294d3c24b4b"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.558710 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" event={"ID":"9a4f1808-03cb-413f-bbfc-01eab48a65d1","Type":"ContainerStarted","Data":"14262c73385158a7e76a4195551145c098d33c68551967b69ee4f25bfd3119ce"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.561937 4667 generic.go:334] "Generic (PLEG): container finished" podID="323592f8-1556-46c0-82d8-0d9fd867ab33" containerID="d00ff2a72d3b7490cbc38cf900509ced1670f40cca16cf4397df9eb79b72f98c" exitCode=0 Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.562097 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" event={"ID":"323592f8-1556-46c0-82d8-0d9fd867ab33","Type":"ContainerDied","Data":"d00ff2a72d3b7490cbc38cf900509ced1670f40cca16cf4397df9eb79b72f98c"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.562136 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" event={"ID":"323592f8-1556-46c0-82d8-0d9fd867ab33","Type":"ContainerStarted","Data":"cb851387797aa205682f904b99b93c8aa0d00e2d2658b2c22670f6c46cbccb82"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.570017 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-b5bmf" event={"ID":"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61","Type":"ContainerStarted","Data":"8825f658d13c3f1bc0908be71c13b1a48a014725ab2f7be7ed5996634519dd96"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.570350 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-b5bmf" event={"ID":"ee9d8867-95b1-4b66-ac9b-3b0f582c8f61","Type":"ContainerStarted","Data":"10fb9074cbaf4a83d37faf5f4e26526828f8a31976a114d8209dad6ce2e1a735"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.570488 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.571389 4667 patch_prober.go:28] interesting pod/console-operator-58897d9998-b5bmf container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.571489 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-b5bmf" podUID="ee9d8867-95b1-4b66-ac9b-3b0f582c8f61" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.574589 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" event={"ID":"d785eccf-d104-4f86-8339-884e0bbc4a52","Type":"ContainerStarted","Data":"563b29829b85d0ee39e3b8423ac6a0fc1d4a0da878f2d1855bfcfa39a3295d75"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.577592 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qhc55" event={"ID":"cab548cd-ee10-421c-9648-02dd9ad58dfa","Type":"ContainerStarted","Data":"8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.577635 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qhc55" event={"ID":"cab548cd-ee10-421c-9648-02dd9ad58dfa","Type":"ContainerStarted","Data":"f369ea26b675f56bf991251edd9e5e94e08698c54213c660939d543c07c7adbc"} Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.612730 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8"] Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.635265 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.637395 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.137382006 +0000 UTC m=+138.635228776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.693688 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vkr48"] Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.736488 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.737682 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.237669363 +0000 UTC m=+138.735516133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.749074 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-tr4hr"] Sep 29 17:11:49 crc kubenswrapper[4667]: W0929 17:11:49.779220 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99797fa8_02d2_42d2_85b6_a96e05a8fe59.slice/crio-dca0586bd3f605582102090331bfd7a3b20448bd4633156c57c2a2f738a08114 WatchSource:0}: Error finding container dca0586bd3f605582102090331bfd7a3b20448bd4633156c57c2a2f738a08114: Status 404 returned error can't find the container with id dca0586bd3f605582102090331bfd7a3b20448bd4633156c57c2a2f738a08114 Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.841097 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.841457 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.341446171 +0000 UTC m=+138.839292941 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:49 crc kubenswrapper[4667]: I0929 17:11:49.945538 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:49 crc kubenswrapper[4667]: E0929 17:11:49.946029 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.446015754 +0000 UTC m=+138.943862523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.051075 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.051369 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.551359063 +0000 UTC m=+139.049205832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.150115 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl"] Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.151977 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.152428 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.652399438 +0000 UTC m=+139.150246206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.159904 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx"] Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.194184 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" podStartSLOduration=119.194168532 podStartE2EDuration="1m59.194168532s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.170010731 +0000 UTC m=+138.667857499" watchObservedRunningTime="2025-09-29 17:11:50.194168532 +0000 UTC m=+138.692015302" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.196249 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7ssxc"] Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.203153 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-qhc55" podStartSLOduration=120.203137767 podStartE2EDuration="2m0.203137767s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.201035196 +0000 UTC m=+138.698881965" watchObservedRunningTime="2025-09-29 17:11:50.203137767 +0000 UTC m=+138.700984535" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.217339 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn"] Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.217371 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj"] Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.227000 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7"] Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.235154 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qn7ls"] Sep 29 17:11:50 crc kubenswrapper[4667]: W0929 17:11:50.236224 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8035cbb_1852_41c8_8aee_820e412f9027.slice/crio-85e159a75dfb072ec99d5ff8d97e49c23c4f6f3d32b8568ba69b7ea82ca96b52 WatchSource:0}: Error finding container 85e159a75dfb072ec99d5ff8d97e49c23c4f6f3d32b8568ba69b7ea82ca96b52: Status 404 returned error can't find the container with id 85e159a75dfb072ec99d5ff8d97e49c23c4f6f3d32b8568ba69b7ea82ca96b52 Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.242035 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-2fnkw" podStartSLOduration=120.242020053 podStartE2EDuration="2m0.242020053s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.241475847 +0000 UTC m=+138.739322617" watchObservedRunningTime="2025-09-29 17:11:50.242020053 +0000 UTC m=+138.739866822" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.255115 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.255395 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.755384035 +0000 UTC m=+139.253230804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.256471 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq"] Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.258953 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wjdtk"] Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.274833 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tz7dc" podStartSLOduration=120.274819403 podStartE2EDuration="2m0.274819403s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.274135514 +0000 UTC m=+138.771982283" watchObservedRunningTime="2025-09-29 17:11:50.274819403 +0000 UTC m=+138.772666172" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.326887 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-b89jg" podStartSLOduration=120.326870964 podStartE2EDuration="2m0.326870964s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.325349708 +0000 UTC m=+138.823196477" watchObservedRunningTime="2025-09-29 17:11:50.326870964 +0000 UTC m=+138.824717733" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.349950 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.355717 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.355810 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.855796737 +0000 UTC m=+139.353643506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.356344 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.356576 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.856568972 +0000 UTC m=+139.354415742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.362276 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tzmmc" podStartSLOduration=119.3622647 podStartE2EDuration="1m59.3622647s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.361706659 +0000 UTC m=+138.859553428" watchObservedRunningTime="2025-09-29 17:11:50.3622647 +0000 UTC m=+138.860111469" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.362622 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:50 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:50 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:50 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.362674 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.374852 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xbcfz"] Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.415759 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lwt9b" podStartSLOduration=120.41574365 podStartE2EDuration="2m0.41574365s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.415005159 +0000 UTC m=+138.912851928" watchObservedRunningTime="2025-09-29 17:11:50.41574365 +0000 UTC m=+138.913590420" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.456669 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-b5bmf" podStartSLOduration=120.456655139 podStartE2EDuration="2m0.456655139s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.455532674 +0000 UTC m=+138.953379444" watchObservedRunningTime="2025-09-29 17:11:50.456655139 +0000 UTC m=+138.954501899" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.458093 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.458440 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:50.958429071 +0000 UTC m=+139.456275840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.477280 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5s6w" podStartSLOduration=120.477262594 podStartE2EDuration="2m0.477262594s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.476590018 +0000 UTC m=+138.974436787" watchObservedRunningTime="2025-09-29 17:11:50.477262594 +0000 UTC m=+138.975109363" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.525259 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-hhstg" podStartSLOduration=5.525247157 podStartE2EDuration="5.525247157s" podCreationTimestamp="2025-09-29 17:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.52420856 +0000 UTC m=+139.022055329" watchObservedRunningTime="2025-09-29 17:11:50.525247157 +0000 UTC m=+139.023093925" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.549832 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fmmvk" podStartSLOduration=120.549818758 podStartE2EDuration="2m0.549818758s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.549281976 +0000 UTC m=+139.047128746" watchObservedRunningTime="2025-09-29 17:11:50.549818758 +0000 UTC m=+139.047665527" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.564723 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.565005 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.064994411 +0000 UTC m=+139.562841181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.589284 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t" event={"ID":"3c761a0b-9b2c-4556-9e1e-060c3529ab38","Type":"ContainerStarted","Data":"9a4a9aecbe9b710b311b198e825d36abde9c80d9db86ac632d75b8258889c6d1"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.589334 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t" event={"ID":"3c761a0b-9b2c-4556-9e1e-060c3529ab38","Type":"ContainerStarted","Data":"4fc8bf511fba4caa4596447333abb7b0cee4456d48ba52362a1fb467f9795733"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.602387 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" event={"ID":"3e396bd3-e718-4f53-a69b-522a601e0d4d","Type":"ContainerStarted","Data":"93b426cd6d4f47c61343774eb9dba5ecea02cff31b9069675bc3fcd8f7854aa3"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.602557 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" event={"ID":"3e396bd3-e718-4f53-a69b-522a601e0d4d","Type":"ContainerStarted","Data":"022bcd7552e7482ee47ef25a7b0f204e0455a41ba85f93d1ed8f81a4e144f1b5"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.631525 4667 generic.go:334] "Generic (PLEG): container finished" podID="c550aca9-a2c1-4251-8a19-e03bd9dada5e" containerID="b22840cba8a1a20fee6e67a62fdffd462a6ef6cc4bb0c22c786d1f915bb7fa63" exitCode=0 Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.631615 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" event={"ID":"c550aca9-a2c1-4251-8a19-e03bd9dada5e","Type":"ContainerDied","Data":"b22840cba8a1a20fee6e67a62fdffd462a6ef6cc4bb0c22c786d1f915bb7fa63"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.636166 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" podStartSLOduration=120.636153333 podStartE2EDuration="2m0.636153333s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.599113316 +0000 UTC m=+139.096960086" watchObservedRunningTime="2025-09-29 17:11:50.636153333 +0000 UTC m=+139.134000102" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.637534 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" event={"ID":"fedae4da-7b65-4159-9248-67f9f09ed3a4","Type":"ContainerStarted","Data":"e60a1996109e7bcb48d6a7831f31202d1143d9cb5737684d0a1896cff845a222"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.640110 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" event={"ID":"8c91ba86-0e55-462b-acf3-08d7308d7df8","Type":"ContainerStarted","Data":"6c2f1694c7bb9a5aced701d57fd8fefa158a2bf69a4f8ac034404386befeef73"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.642142 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" event={"ID":"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd","Type":"ContainerStarted","Data":"16a23927eaccf40a0b19f9d0ade55a022d79369ed277af5640bebab844543898"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.645862 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" event={"ID":"56908f94-88ac-4298-98e5-3ba1315bd4ae","Type":"ContainerStarted","Data":"175d85e12165607e664750dc01b891ce89fde0e0fcd07ea3ccceac90424bb52f"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.648639 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xbcfz" event={"ID":"d31bb6f3-436e-47fe-b32a-be422ad9825b","Type":"ContainerStarted","Data":"a6d79cdd9d351d455563e8938b7e1da471589f42b86e5f18db4812dc9a56d7ce"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.651288 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" event={"ID":"e32bf925-6e3f-4c59-bbb9-1b079a73163d","Type":"ContainerStarted","Data":"58c17f97b0b34e2f5be093314dae61b3ec4d59e63a38b2dd874ec02c224aafd1"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.651326 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" event={"ID":"e32bf925-6e3f-4c59-bbb9-1b079a73163d","Type":"ContainerStarted","Data":"291f3410a43ce0a3422a5375c9bd260a6fe7a53d6f5ea744947471d4963fc9be"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.651341 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.653441 4667 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dg5gl container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" start-of-body= Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.653492 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" podUID="e32bf925-6e3f-4c59-bbb9-1b079a73163d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.33:5443/healthz\": dial tcp 10.217.0.33:5443: connect: connection refused" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.661837 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hhstg" event={"ID":"91772b5e-ad6f-4830-8152-4e63306b448a","Type":"ContainerStarted","Data":"a7d2b828dfe8a3c720942e21c5f7d3aa66f0c24cf8392a85f844c1a3ac28214f"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.665096 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" event={"ID":"55f01238-26fc-4f87-a22b-19024c896023","Type":"ContainerStarted","Data":"6df692162f0303f48d9bbbb23ad272e79bcd554b0f6167c3b18e4c5ad5807ea7"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.666378 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.666501 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.166487118 +0000 UTC m=+139.664333888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.667178 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.669499 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.169487019 +0000 UTC m=+139.667333789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.680261 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" event={"ID":"62a04aa6-d927-4890-b405-c4dfea426f50","Type":"ContainerStarted","Data":"23dd3ae7c829e7f3d0b886dac8c5b571eab7120bf69c0f92dafbf0ca322f73c1"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.680292 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" event={"ID":"62a04aa6-d927-4890-b405-c4dfea426f50","Type":"ContainerStarted","Data":"fd76e219bf0502afb3d69519915862acc9287acfaae147d0142958fbed3db129"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.683704 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-fflcd" podStartSLOduration=120.683688799 podStartE2EDuration="2m0.683688799s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.679721877 +0000 UTC m=+139.177568636" watchObservedRunningTime="2025-09-29 17:11:50.683688799 +0000 UTC m=+139.181535568" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.696560 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" event={"ID":"9a4f1808-03cb-413f-bbfc-01eab48a65d1","Type":"ContainerStarted","Data":"0fc8a1882649de97c32ea8bb85a166778d73c68976fb6045aede92ad3d7b2514"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.697926 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.707906 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" event={"ID":"c8035cbb-1852-41c8-8aee-820e412f9027","Type":"ContainerStarted","Data":"85e159a75dfb072ec99d5ff8d97e49c23c4f6f3d32b8568ba69b7ea82ca96b52"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.718299 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.724451 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" event={"ID":"323592f8-1556-46c0-82d8-0d9fd867ab33","Type":"ContainerStarted","Data":"142642966a8c2fafbcec3522bc8e4365e39d4bf5bd29f31cfb870142f39f5045"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.729058 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b89jg" event={"ID":"e0752d76-2a38-4092-ae1e-d5a9d4b395e1","Type":"ContainerStarted","Data":"91464f6c088ae34b3c3a7b74e3c6204592ef64ea4199ec318b9ac4a893b0380b"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.729798 4667 patch_prober.go:28] interesting pod/downloads-7954f5f757-b89jg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.729822 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b89jg" podUID="e0752d76-2a38-4092-ae1e-d5a9d4b395e1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.770540 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.771159 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.271143284 +0000 UTC m=+139.768990053 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.772601 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" event={"ID":"8e68faca-8329-4201-a26d-e90968c0a8fb","Type":"ContainerStarted","Data":"3f9ac732f05bae33866b0d3a96ca8bb02eb02ccfcb6c09d044ebd0ecf20e9f58"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.772638 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" event={"ID":"8e68faca-8329-4201-a26d-e90968c0a8fb","Type":"ContainerStarted","Data":"309d5cb15d52abda86c99af078a2ef3fee2b61c65cea7099496df62f9433011c"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.793488 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" event={"ID":"99797fa8-02d2-42d2-85b6-a96e05a8fe59","Type":"ContainerStarted","Data":"3225d2e933ed020ec2e5ff5880f54eb5e506dcba28c82c974c90e93cdb937cb5"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.793526 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" event={"ID":"99797fa8-02d2-42d2-85b6-a96e05a8fe59","Type":"ContainerStarted","Data":"dca0586bd3f605582102090331bfd7a3b20448bd4633156c57c2a2f738a08114"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.811531 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2szrg" podStartSLOduration=120.81151706 podStartE2EDuration="2m0.81151706s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.759534529 +0000 UTC m=+139.257381297" watchObservedRunningTime="2025-09-29 17:11:50.81151706 +0000 UTC m=+139.309363829" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.817129 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" event={"ID":"9807d262-d148-47f0-afd5-9b92aacfba38","Type":"ContainerStarted","Data":"3fe7830a7842f435ab64ef2b7fcfe4eb83ba8422c286f56fc90c974eea0c4126"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.817894 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.832057 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.853076 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" event={"ID":"b20d92fc-2533-47f3-a553-7285c85ca8d0","Type":"ContainerStarted","Data":"498d9daa5571a33e62de6031d7e66c4411594a49dfaf6a129f8a57de020da16e"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.853118 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" event={"ID":"b20d92fc-2533-47f3-a553-7285c85ca8d0","Type":"ContainerStarted","Data":"1ab1f69323f79e0cc17718e3f658d6db197714a6c1b3a8c2e16a8eb3643e4922"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.854015 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" podStartSLOduration=120.854004396 podStartE2EDuration="2m0.854004396s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.813186694 +0000 UTC m=+139.311033464" watchObservedRunningTime="2025-09-29 17:11:50.854004396 +0000 UTC m=+139.351851166" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.872698 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" event={"ID":"37d1b887-4e1d-4ea3-b247-3c6091047751","Type":"ContainerStarted","Data":"45afc36f8847ee720263df7a11b5005df298cacc43eb3ad8bc5f171b880237de"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.872735 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" event={"ID":"37d1b887-4e1d-4ea3-b247-3c6091047751","Type":"ContainerStarted","Data":"6df822d4f7d74ebdbb485f429fc600123359c2f1cc4f6a9b66635d2eef8d29d3"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.882352 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.885173 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.38516003 +0000 UTC m=+139.883006799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.894635 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2kwqd" podStartSLOduration=119.894622734 podStartE2EDuration="1m59.894622734s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.89406866 +0000 UTC m=+139.391915430" watchObservedRunningTime="2025-09-29 17:11:50.894622734 +0000 UTC m=+139.392469502" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.898135 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" event={"ID":"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2","Type":"ContainerStarted","Data":"43aea7ce1aff1605db00fce3ce3bb989b0242f18c2ff3dfd831faba070482a56"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.898177 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" event={"ID":"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2","Type":"ContainerStarted","Data":"08b4de814a7d084b04508e672c7e361b202fd222b3411a5241a96459484d2c6c"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.920592 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pk54t" podStartSLOduration=120.920574324 podStartE2EDuration="2m0.920574324s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:50.918165899 +0000 UTC m=+139.416012668" watchObservedRunningTime="2025-09-29 17:11:50.920574324 +0000 UTC m=+139.418421094" Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.931640 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" event={"ID":"d785eccf-d104-4f86-8339-884e0bbc4a52","Type":"ContainerStarted","Data":"d9b5c4d43509ad553f31235da2e2e1b676288e223e4378a579692f1c69c896df"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.931674 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" event={"ID":"d785eccf-d104-4f86-8339-884e0bbc4a52","Type":"ContainerStarted","Data":"057321bd97e5c93d6004d38ff889749094d1b94372a8f13a4300fe708ea82def"} Sep 29 17:11:50 crc kubenswrapper[4667]: I0929 17:11:50.988335 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:50 crc kubenswrapper[4667]: E0929 17:11:50.993861 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.488892405 +0000 UTC m=+139.986739174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.014137 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-q78dk" podStartSLOduration=121.014122668 podStartE2EDuration="2m1.014122668s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.013458757 +0000 UTC m=+139.511305526" watchObservedRunningTime="2025-09-29 17:11:51.014122668 +0000 UTC m=+139.511969437" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.028046 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" event={"ID":"5c692723-1438-4ec7-8e1a-d521d3f6e981","Type":"ContainerStarted","Data":"fbb762b89feeb6bf87e05ede7833fa1bda01d2a0859927aacf12e628c3ab3bae"} Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.071105 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" event={"ID":"58bab54c-7710-4347-a85f-9b1554683c06","Type":"ContainerStarted","Data":"faadb8b7d1d394fafb2dd4d8fcb9bef7dc01257a2bf40e74bb241366dd8ec3c9"} Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.072035 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.084806 4667 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-p9kvq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.084856 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" podUID="58bab54c-7710-4347-a85f-9b1554683c06" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.092431 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:51 crc kubenswrapper[4667]: E0929 17:11:51.095497 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.595483786 +0000 UTC m=+140.093330555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.101276 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tr4hr" event={"ID":"624ad822-49b1-4cdc-851b-5727b8fdbf91","Type":"ContainerStarted","Data":"95c6e5d611f000bc2901e84c5935a31e0f3d814d332514aa8a9da92ef72b54db"} Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.101319 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-tr4hr" event={"ID":"624ad822-49b1-4cdc-851b-5727b8fdbf91","Type":"ContainerStarted","Data":"4aeaa0c8efbb611f9ce763058e55179625ab47c6584ca882edc636c2f2490786"} Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.118049 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2m7w5" podStartSLOduration=121.118036115 podStartE2EDuration="2m1.118036115s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.084595047 +0000 UTC m=+139.582441816" watchObservedRunningTime="2025-09-29 17:11:51.118036115 +0000 UTC m=+139.615882873" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.120084 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" event={"ID":"ee052316-056b-4cb5-b05c-b554502091da","Type":"ContainerStarted","Data":"9f4efb03e0126df53919b1c4d195bf6ce749675d628824323213da34a8ad8ebc"} Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.120198 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" event={"ID":"ee052316-056b-4cb5-b05c-b554502091da","Type":"ContainerStarted","Data":"1f9a6623afb0f04156f00a5e7eda66b4a3953e08e61ba3f529fac944b216e0d6"} Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.132683 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-b5bmf" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.140331 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.146652 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gtjfr" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.162218 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5bpkn" podStartSLOduration=120.162202754 podStartE2EDuration="2m0.162202754s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.132999569 +0000 UTC m=+139.630846337" watchObservedRunningTime="2025-09-29 17:11:51.162202754 +0000 UTC m=+139.660049523" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.176352 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-2npg8" podStartSLOduration=121.176340003 podStartE2EDuration="2m1.176340003s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.17548391 +0000 UTC m=+139.673330679" watchObservedRunningTime="2025-09-29 17:11:51.176340003 +0000 UTC m=+139.674186771" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.193141 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:51 crc kubenswrapper[4667]: E0929 17:11:51.194046 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.694032648 +0000 UTC m=+140.191879416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.196806 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-7x9v7" podStartSLOduration=120.196796603 podStartE2EDuration="2m0.196796603s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.19509572 +0000 UTC m=+139.692942489" watchObservedRunningTime="2025-09-29 17:11:51.196796603 +0000 UTC m=+139.694643372" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.257613 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" podStartSLOduration=120.257599118 podStartE2EDuration="2m0.257599118s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.230172679 +0000 UTC m=+139.728019448" watchObservedRunningTime="2025-09-29 17:11:51.257599118 +0000 UTC m=+139.755445887" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.293249 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" podStartSLOduration=120.293235381 podStartE2EDuration="2m0.293235381s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.25843905 +0000 UTC m=+139.756285819" watchObservedRunningTime="2025-09-29 17:11:51.293235381 +0000 UTC m=+139.791082150" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.294000 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c2zxb" podStartSLOduration=122.293995803 podStartE2EDuration="2m2.293995803s" podCreationTimestamp="2025-09-29 17:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.292817614 +0000 UTC m=+139.790664383" watchObservedRunningTime="2025-09-29 17:11:51.293995803 +0000 UTC m=+139.791842572" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.295714 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:51 crc kubenswrapper[4667]: E0929 17:11:51.297822 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.79781133 +0000 UTC m=+140.295658099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.312742 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-lbx5r" podStartSLOduration=120.312727506 podStartE2EDuration="2m0.312727506s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.31059054 +0000 UTC m=+139.808437310" watchObservedRunningTime="2025-09-29 17:11:51.312727506 +0000 UTC m=+139.810574264" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.360346 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" podStartSLOduration=121.360317803 podStartE2EDuration="2m1.360317803s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.35993321 +0000 UTC m=+139.857779978" watchObservedRunningTime="2025-09-29 17:11:51.360317803 +0000 UTC m=+139.858164572" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.362006 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" podStartSLOduration=121.361997408 podStartE2EDuration="2m1.361997408s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.330359837 +0000 UTC m=+139.828206606" watchObservedRunningTime="2025-09-29 17:11:51.361997408 +0000 UTC m=+139.859844177" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.362479 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:51 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:51 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:51 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.362509 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.404648 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:51 crc kubenswrapper[4667]: E0929 17:11:51.404977 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:51.904964578 +0000 UTC m=+140.402811347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.485668 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" podStartSLOduration=120.485652729 podStartE2EDuration="2m0.485652729s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.485524728 +0000 UTC m=+139.983371496" watchObservedRunningTime="2025-09-29 17:11:51.485652729 +0000 UTC m=+139.983499497" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.506607 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:51 crc kubenswrapper[4667]: E0929 17:11:51.506862 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.006837552 +0000 UTC m=+140.504684320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.573663 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vb7qx" podStartSLOduration=120.573649455 podStartE2EDuration="2m0.573649455s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.558094988 +0000 UTC m=+140.055941757" watchObservedRunningTime="2025-09-29 17:11:51.573649455 +0000 UTC m=+140.071496224" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.609100 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:51 crc kubenswrapper[4667]: E0929 17:11:51.609489 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.109468192 +0000 UTC m=+140.607314961 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.710100 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:51 crc kubenswrapper[4667]: E0929 17:11:51.710589 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.210578768 +0000 UTC m=+140.708425527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.793767 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-tr4hr" podStartSLOduration=6.793747831 podStartE2EDuration="6.793747831s" podCreationTimestamp="2025-09-29 17:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:51.607132743 +0000 UTC m=+140.104979511" watchObservedRunningTime="2025-09-29 17:11:51.793747831 +0000 UTC m=+140.291594600" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.795212 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z7rjk"] Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.803584 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.812044 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.812321 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z7rjk"] Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.812580 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 29 17:11:51 crc kubenswrapper[4667]: E0929 17:11:51.812749 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.312734013 +0000 UTC m=+140.810580782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.914372 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-utilities\") pod \"certified-operators-z7rjk\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.914434 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-catalog-content\") pod \"certified-operators-z7rjk\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.914480 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chwt7\" (UniqueName: \"kubernetes.io/projected/3795f5d0-58d5-40a1-bfcb-98da8a26e905-kube-api-access-chwt7\") pod \"certified-operators-z7rjk\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.914581 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:51 crc kubenswrapper[4667]: E0929 17:11:51.914980 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.414962556 +0000 UTC m=+140.912809325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.984680 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n26jh"] Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.985576 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:51 crc kubenswrapper[4667]: I0929 17:11:51.989730 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.014601 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n26jh"] Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.014969 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.015287 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-utilities\") pod \"certified-operators-z7rjk\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.015322 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-catalog-content\") pod \"certified-operators-z7rjk\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.015353 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chwt7\" (UniqueName: \"kubernetes.io/projected/3795f5d0-58d5-40a1-bfcb-98da8a26e905-kube-api-access-chwt7\") pod \"certified-operators-z7rjk\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.016444 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-utilities\") pod \"certified-operators-z7rjk\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.016534 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.516520154 +0000 UTC m=+141.014366923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.024032 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-catalog-content\") pod \"certified-operators-z7rjk\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.060092 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chwt7\" (UniqueName: \"kubernetes.io/projected/3795f5d0-58d5-40a1-bfcb-98da8a26e905-kube-api-access-chwt7\") pod \"certified-operators-z7rjk\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.116577 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.116858 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-utilities\") pod \"community-operators-n26jh\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.116980 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w67n5\" (UniqueName: \"kubernetes.io/projected/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-kube-api-access-w67n5\") pod \"community-operators-n26jh\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.117139 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-catalog-content\") pod \"community-operators-n26jh\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.117510 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.617499504 +0000 UTC m=+141.115346273 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.130084 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.134873 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" event={"ID":"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd","Type":"ContainerStarted","Data":"a5574791af3262c81e35b8b1f458dfabf1aea27a7e8c754264c85a81c82666ce"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.153367 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xbcfz" event={"ID":"d31bb6f3-436e-47fe-b32a-be422ad9825b","Type":"ContainerStarted","Data":"32f069f4fa1924a8956ab37612305ecf98a12a587b2c2e857d69948bb05e2443"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.153415 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xbcfz" event={"ID":"d31bb6f3-436e-47fe-b32a-be422ad9825b","Type":"ContainerStarted","Data":"6523b7c0a36bf887c67270ac5b6574a3911e15600d0e091386c7f14166286dc8"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.154144 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-xbcfz" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.162899 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" event={"ID":"c8035cbb-1852-41c8-8aee-820e412f9027","Type":"ContainerStarted","Data":"1eb775fd6d2ffd51314a45e9d5d6a7aff1827ba980be20e9289148ba95db811e"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.162924 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" event={"ID":"c8035cbb-1852-41c8-8aee-820e412f9027","Type":"ContainerStarted","Data":"80d07cbdd152c559ba1c62693505ec8f430a8fe390a884dd1ff390e24a8c40f3"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.169604 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" event={"ID":"c550aca9-a2c1-4251-8a19-e03bd9dada5e","Type":"ContainerStarted","Data":"2bdf449be688d5700625bc61488fc408162a82010170ea7a775bf751dc081d74"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.169629 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" event={"ID":"c550aca9-a2c1-4251-8a19-e03bd9dada5e","Type":"ContainerStarted","Data":"2f7defbdad9a5516c9768c1f6886836b9c4dd4e09c159a8311e34466e8d77c91"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.171350 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" event={"ID":"37d1b887-4e1d-4ea3-b247-3c6091047751","Type":"ContainerStarted","Data":"7ce379f74344cdd574f0cdeb1688c9c61895cf361e1ba06b381115f5cde7d4bd"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.174300 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" event={"ID":"56908f94-88ac-4298-98e5-3ba1315bd4ae","Type":"ContainerStarted","Data":"d6efad009de94276a9d0733d1f15f31a8bf612324fb895731346587f8e2d3fad"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.174933 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.181576 4667 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wjdtk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.181712 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" podUID="56908f94-88ac-4298-98e5-3ba1315bd4ae" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.182970 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-xbcfz" podStartSLOduration=7.182960764 podStartE2EDuration="7.182960764s" podCreationTimestamp="2025-09-29 17:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:52.175168236 +0000 UTC m=+140.673015005" watchObservedRunningTime="2025-09-29 17:11:52.182960764 +0000 UTC m=+140.680807533" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.186247 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5t5jw"] Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.186969 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.189891 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5t5jw"] Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.192911 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" event={"ID":"5c692723-1438-4ec7-8e1a-d521d3f6e981","Type":"ContainerStarted","Data":"77a39cab22120efc1bf3736e5fc2464b4679674418ee0da020ece868de09576a"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.192938 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" event={"ID":"5c692723-1438-4ec7-8e1a-d521d3f6e981","Type":"ContainerStarted","Data":"e97d7b56690223d622a1566f3a8231410e696b32125645c01265d5f4ca1e8b3c"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.193343 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.197233 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" event={"ID":"fedae4da-7b65-4159-9248-67f9f09ed3a4","Type":"ContainerStarted","Data":"0eee7d8b1a9a018db41ada2315ae1faf15f4b110882c7b07e00d0ddda6697eaf"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.201595 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" event={"ID":"58bab54c-7710-4347-a85f-9b1554683c06","Type":"ContainerStarted","Data":"12c4e9c813d4f64852f7a32b0f59460e8877d30ea895daeda096a48259af1a2e"} Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.202302 4667 patch_prober.go:28] interesting pod/downloads-7954f5f757-b89jg container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.202341 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b89jg" podUID="e0752d76-2a38-4092-ae1e-d5a9d4b395e1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.229395 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.229829 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-catalog-content\") pod \"community-operators-n26jh\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.229971 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-utilities\") pod \"community-operators-n26jh\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.229999 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w67n5\" (UniqueName: \"kubernetes.io/projected/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-kube-api-access-w67n5\") pod \"community-operators-n26jh\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.230544 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.730524894 +0000 UTC m=+141.228371662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.230613 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-7ssxc" podStartSLOduration=122.230598111 podStartE2EDuration="2m2.230598111s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:52.208604345 +0000 UTC m=+140.706451114" watchObservedRunningTime="2025-09-29 17:11:52.230598111 +0000 UTC m=+140.728445040" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.232793 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-p9kvq" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.233195 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-catalog-content\") pod \"community-operators-n26jh\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.237107 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-utilities\") pod \"community-operators-n26jh\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.254338 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-vkr48" podStartSLOduration=121.254325342 podStartE2EDuration="2m1.254325342s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:52.251543894 +0000 UTC m=+140.749390663" watchObservedRunningTime="2025-09-29 17:11:52.254325342 +0000 UTC m=+140.752172112" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.283745 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w67n5\" (UniqueName: \"kubernetes.io/projected/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-kube-api-access-w67n5\") pod \"community-operators-n26jh\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.285417 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" podStartSLOduration=121.285392539 podStartE2EDuration="2m1.285392539s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:52.283964539 +0000 UTC m=+140.781811308" watchObservedRunningTime="2025-09-29 17:11:52.285392539 +0000 UTC m=+140.783239308" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.301182 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.332348 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-utilities\") pod \"certified-operators-5t5jw\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.332534 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.332919 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn92v\" (UniqueName: \"kubernetes.io/projected/dcbcc4f2-3dff-49dc-b310-8504021366d3-kube-api-access-qn92v\") pod \"certified-operators-5t5jw\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.332973 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-catalog-content\") pod \"certified-operators-5t5jw\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.346824 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.846810253 +0000 UTC m=+141.344657022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.357425 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" podStartSLOduration=122.357398256 podStartE2EDuration="2m2.357398256s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:52.327513627 +0000 UTC m=+140.825360396" watchObservedRunningTime="2025-09-29 17:11:52.357398256 +0000 UTC m=+140.855245025" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.380762 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:52 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:52 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:52 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.380811 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.391618 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rh4rn" podStartSLOduration=121.391600627 podStartE2EDuration="2m1.391600627s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:52.389294505 +0000 UTC m=+140.887141273" watchObservedRunningTime="2025-09-29 17:11:52.391600627 +0000 UTC m=+140.889447397" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.414142 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2s5vw"] Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.415094 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2s5vw"] Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.415181 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.437975 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.438216 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-utilities\") pod \"certified-operators-5t5jw\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.438340 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn92v\" (UniqueName: \"kubernetes.io/projected/dcbcc4f2-3dff-49dc-b310-8504021366d3-kube-api-access-qn92v\") pod \"certified-operators-5t5jw\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.438369 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-catalog-content\") pod \"certified-operators-5t5jw\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.439551 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-catalog-content\") pod \"certified-operators-5t5jw\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.439796 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-utilities\") pod \"certified-operators-5t5jw\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.439990 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:52.939964283 +0000 UTC m=+141.437811053 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.464529 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn92v\" (UniqueName: \"kubernetes.io/projected/dcbcc4f2-3dff-49dc-b310-8504021366d3-kube-api-access-qn92v\") pod \"certified-operators-5t5jw\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.472330 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" podStartSLOduration=121.472312793 podStartE2EDuration="2m1.472312793s" podCreationTimestamp="2025-09-29 17:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:52.472179922 +0000 UTC m=+140.970026692" watchObservedRunningTime="2025-09-29 17:11:52.472312793 +0000 UTC m=+140.970159561" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.503138 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.539685 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-catalog-content\") pod \"community-operators-2s5vw\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.539727 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-utilities\") pod \"community-operators-2s5vw\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.539775 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.539901 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvt94\" (UniqueName: \"kubernetes.io/projected/28489e28-fd28-4526-99dd-4f72aefbc272-kube-api-access-fvt94\") pod \"community-operators-2s5vw\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.540247 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.04023135 +0000 UTC m=+141.538078120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.629401 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z7rjk"] Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.641013 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.641120 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.141096686 +0000 UTC m=+141.638943455 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.641337 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvt94\" (UniqueName: \"kubernetes.io/projected/28489e28-fd28-4526-99dd-4f72aefbc272-kube-api-access-fvt94\") pod \"community-operators-2s5vw\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.641387 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-catalog-content\") pod \"community-operators-2s5vw\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.641427 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-utilities\") pod \"community-operators-2s5vw\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.641464 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.641827 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-catalog-content\") pod \"community-operators-2s5vw\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.641878 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.14186298 +0000 UTC m=+141.639709749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.641910 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-utilities\") pod \"community-operators-2s5vw\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.685442 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvt94\" (UniqueName: \"kubernetes.io/projected/28489e28-fd28-4526-99dd-4f72aefbc272-kube-api-access-fvt94\") pod \"community-operators-2s5vw\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.746229 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.746869 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.246836723 +0000 UTC m=+141.744683492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.793652 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n26jh"] Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.847781 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.848104 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.348093966 +0000 UTC m=+141.845940735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.851198 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.851252 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.858792 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dg5gl" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.869268 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.893077 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.948493 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.948769 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.448746831 +0000 UTC m=+141.946593600 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:52 crc kubenswrapper[4667]: I0929 17:11:52.949614 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:52 crc kubenswrapper[4667]: E0929 17:11:52.950151 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.450044986 +0000 UTC m=+141.947891754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.029125 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.029557 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.049926 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5t5jw"] Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.053262 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.053541 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.553527861 +0000 UTC m=+142.051374630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.154953 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.155175 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.655165641 +0000 UTC m=+142.153012411 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.205987 4667 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.228448 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5t5jw" event={"ID":"dcbcc4f2-3dff-49dc-b310-8504021366d3","Type":"ContainerStarted","Data":"0f2373ee6e55b86dee82dd899d123f32c95ebeec0c18e6f5cb61f5a9d49b3d41"} Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.231159 4667 generic.go:334] "Generic (PLEG): container finished" podID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerID="a626d823b2ec37753ecfaf1d3789049003e1bb7ad49a95702259a1c33a637049" exitCode=0 Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.231212 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7rjk" event={"ID":"3795f5d0-58d5-40a1-bfcb-98da8a26e905","Type":"ContainerDied","Data":"a626d823b2ec37753ecfaf1d3789049003e1bb7ad49a95702259a1c33a637049"} Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.231232 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7rjk" event={"ID":"3795f5d0-58d5-40a1-bfcb-98da8a26e905","Type":"ContainerStarted","Data":"813b59d342db02717ed58b05a4e6f01b8b30d556e3c8eda487878fa9eecec532"} Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.232721 4667 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.234244 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" event={"ID":"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd","Type":"ContainerStarted","Data":"e41b2c4432ff38a7bf9cf9059b49372ddd0aa0927958fb17b325acc2470d7872"} Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.234276 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" event={"ID":"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd","Type":"ContainerStarted","Data":"f70f2d732b615c625865ac97add932f00b6d7eb215654201c8de8dec8078c12f"} Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.236298 4667 generic.go:334] "Generic (PLEG): container finished" podID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerID="7d7b6840ebde857f7144cdbb1b87e57d10422b41a1137a397ae73180f651c098" exitCode=0 Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.237730 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n26jh" event={"ID":"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b","Type":"ContainerDied","Data":"7d7b6840ebde857f7144cdbb1b87e57d10422b41a1137a397ae73180f651c098"} Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.237752 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n26jh" event={"ID":"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b","Type":"ContainerStarted","Data":"0fe4580fb39db149701f2684e926f1c8e9889df65d6bf0fb42b51c245504300d"} Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.244741 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.245011 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-pk8c9" Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.255991 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.256513 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.75650028 +0000 UTC m=+142.254347050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.296143 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2s5vw"] Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.354873 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:53 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:53 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:53 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.354917 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.361726 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.362733 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.862722766 +0000 UTC m=+142.360569535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.463373 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.463481 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.963466672 +0000 UTC m=+142.461313441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.463631 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.463885 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:53.963878298 +0000 UTC m=+142.461725066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.565153 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.565486 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:54.065473307 +0000 UTC m=+142.563320076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.666814 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.667133 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:54.16711827 +0000 UTC m=+142.664965038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.768079 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.768227 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:54.26820945 +0000 UTC m=+142.766056220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.768327 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.768588 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:54.26858098 +0000 UTC m=+142.766427739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.842252 4667 patch_prober.go:28] interesting pod/apiserver-76f77b778f-rj4qf container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Sep 29 17:11:53 crc kubenswrapper[4667]: [+]log ok Sep 29 17:11:53 crc kubenswrapper[4667]: [+]etcd ok Sep 29 17:11:53 crc kubenswrapper[4667]: [+]poststarthook/start-apiserver-admission-initializer ok Sep 29 17:11:53 crc kubenswrapper[4667]: [+]poststarthook/generic-apiserver-start-informers ok Sep 29 17:11:53 crc kubenswrapper[4667]: [+]poststarthook/max-in-flight-filter ok Sep 29 17:11:53 crc kubenswrapper[4667]: [+]poststarthook/storage-object-count-tracker-hook ok Sep 29 17:11:53 crc kubenswrapper[4667]: [+]poststarthook/image.openshift.io-apiserver-caches ok Sep 29 17:11:53 crc kubenswrapper[4667]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Sep 29 17:11:53 crc kubenswrapper[4667]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Sep 29 17:11:53 crc kubenswrapper[4667]: [+]poststarthook/project.openshift.io-projectcache ok Sep 29 17:11:53 crc kubenswrapper[4667]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Sep 29 17:11:53 crc kubenswrapper[4667]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Sep 29 17:11:53 crc kubenswrapper[4667]: [+]poststarthook/openshift.io-restmapperupdater ok Sep 29 17:11:53 crc kubenswrapper[4667]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Sep 29 17:11:53 crc kubenswrapper[4667]: livez check failed Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.842299 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" podUID="c550aca9-a2c1-4251-8a19-e03bd9dada5e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.869657 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.869759 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-09-29 17:11:54.369743495 +0000 UTC m=+142.867590265 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.869939 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:53 crc kubenswrapper[4667]: E0929 17:11:53.870215 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-09-29 17:11:54.37020787 +0000 UTC m=+142.868054639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-f44b9" (UID: "f2261944-710e-493f-9360-de4dea650ae5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.948826 4667 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-09-29T17:11:53.20601024Z","Handler":null,"Name":""} Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.952764 4667 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.952799 4667 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.971820 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.977778 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5w8pd"] Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.978638 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.980078 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.988283 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w8pd"] Sep 29 17:11:53 crc kubenswrapper[4667]: I0929 17:11:53.993250 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.072633 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.072700 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-utilities\") pod \"redhat-marketplace-5w8pd\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.072831 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjxhv\" (UniqueName: \"kubernetes.io/projected/3acdff5b-96b2-4100-a595-3c17295f9ef6-kube-api-access-rjxhv\") pod \"redhat-marketplace-5w8pd\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.072881 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-catalog-content\") pod \"redhat-marketplace-5w8pd\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.074753 4667 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.074788 4667 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.088824 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-f44b9\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.174446 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjxhv\" (UniqueName: \"kubernetes.io/projected/3acdff5b-96b2-4100-a595-3c17295f9ef6-kube-api-access-rjxhv\") pod \"redhat-marketplace-5w8pd\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.174833 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-catalog-content\") pod \"redhat-marketplace-5w8pd\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.174882 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-utilities\") pod \"redhat-marketplace-5w8pd\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.175255 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-catalog-content\") pod \"redhat-marketplace-5w8pd\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.175298 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-utilities\") pod \"redhat-marketplace-5w8pd\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.191306 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjxhv\" (UniqueName: \"kubernetes.io/projected/3acdff5b-96b2-4100-a595-3c17295f9ef6-kube-api-access-rjxhv\") pod \"redhat-marketplace-5w8pd\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.246812 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" event={"ID":"8b8e46ca-78c8-4c38-8376-9e4d2c0b1edd","Type":"ContainerStarted","Data":"bd60c49e5993133d47b307dafbc0a7884fd177008203d064e43e315527411927"} Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.250217 4667 generic.go:334] "Generic (PLEG): container finished" podID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerID="21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd" exitCode=0 Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.250300 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5t5jw" event={"ID":"dcbcc4f2-3dff-49dc-b310-8504021366d3","Type":"ContainerDied","Data":"21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd"} Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.256706 4667 generic.go:334] "Generic (PLEG): container finished" podID="68e21fe0-a72c-42e2-b2a8-89f8eed86bd2" containerID="43aea7ce1aff1605db00fce3ce3bb989b0242f18c2ff3dfd831faba070482a56" exitCode=0 Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.256775 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" event={"ID":"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2","Type":"ContainerDied","Data":"43aea7ce1aff1605db00fce3ce3bb989b0242f18c2ff3dfd831faba070482a56"} Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.260987 4667 generic.go:334] "Generic (PLEG): container finished" podID="28489e28-fd28-4526-99dd-4f72aefbc272" containerID="6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285" exitCode=0 Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.261226 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2s5vw" event={"ID":"28489e28-fd28-4526-99dd-4f72aefbc272","Type":"ContainerDied","Data":"6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285"} Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.261270 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2s5vw" event={"ID":"28489e28-fd28-4526-99dd-4f72aefbc272","Type":"ContainerStarted","Data":"5840676216b8c371277806c29734544fefb27a895baddeeedeb4d425e8ff46fc"} Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.264527 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-qn7ls" podStartSLOduration=9.264514035 podStartE2EDuration="9.264514035s" podCreationTimestamp="2025-09-29 17:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:54.263015062 +0000 UTC m=+142.760861831" watchObservedRunningTime="2025-09-29 17:11:54.264514035 +0000 UTC m=+142.762360804" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.282512 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.290232 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.350824 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:54 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:54 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:54 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.350895 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.380762 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-62g78"] Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.382508 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.399140 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-62g78"] Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.482722 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgxxk\" (UniqueName: \"kubernetes.io/projected/cb52efdd-e2d7-493e-b786-34eb5e260d40-kube-api-access-hgxxk\") pod \"redhat-marketplace-62g78\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.483021 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-catalog-content\") pod \"redhat-marketplace-62g78\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.483069 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-utilities\") pod \"redhat-marketplace-62g78\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.506952 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f44b9"] Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.539928 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w8pd"] Sep 29 17:11:54 crc kubenswrapper[4667]: W0929 17:11:54.548203 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3acdff5b_96b2_4100_a595_3c17295f9ef6.slice/crio-399a0425620f55ed940d7f309ebab2766c5ba5bc2b2a4b1e4dff06cb61635b24 WatchSource:0}: Error finding container 399a0425620f55ed940d7f309ebab2766c5ba5bc2b2a4b1e4dff06cb61635b24: Status 404 returned error can't find the container with id 399a0425620f55ed940d7f309ebab2766c5ba5bc2b2a4b1e4dff06cb61635b24 Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.584149 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgxxk\" (UniqueName: \"kubernetes.io/projected/cb52efdd-e2d7-493e-b786-34eb5e260d40-kube-api-access-hgxxk\") pod \"redhat-marketplace-62g78\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.584236 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-catalog-content\") pod \"redhat-marketplace-62g78\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.584618 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-catalog-content\") pod \"redhat-marketplace-62g78\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.584670 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-utilities\") pod \"redhat-marketplace-62g78\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.584990 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-utilities\") pod \"redhat-marketplace-62g78\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.598625 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgxxk\" (UniqueName: \"kubernetes.io/projected/cb52efdd-e2d7-493e-b786-34eb5e260d40-kube-api-access-hgxxk\") pod \"redhat-marketplace-62g78\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.706833 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.863969 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-62g78"] Sep 29 17:11:54 crc kubenswrapper[4667]: W0929 17:11:54.878897 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb52efdd_e2d7_493e_b786_34eb5e260d40.slice/crio-af36384da35e347f41f935c8fdd6f4151388f0f96f4de798a6440b35b1ab538d WatchSource:0}: Error finding container af36384da35e347f41f935c8fdd6f4151388f0f96f4de798a6440b35b1ab538d: Status 404 returned error can't find the container with id af36384da35e347f41f935c8fdd6f4151388f0f96f4de798a6440b35b1ab538d Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.975600 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2skv5"] Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.977208 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.978718 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 29 17:11:54 crc kubenswrapper[4667]: I0929 17:11:54.987981 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2skv5"] Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.090872 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-catalog-content\") pod \"redhat-operators-2skv5\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.090994 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-utilities\") pod \"redhat-operators-2skv5\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.091047 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxll9\" (UniqueName: \"kubernetes.io/projected/af6b10ff-6922-44fc-a489-514cbcffe69d-kube-api-access-vxll9\") pod \"redhat-operators-2skv5\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.156654 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.158109 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.160097 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.160346 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.164781 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.191875 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-utilities\") pod \"redhat-operators-2skv5\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.191936 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxll9\" (UniqueName: \"kubernetes.io/projected/af6b10ff-6922-44fc-a489-514cbcffe69d-kube-api-access-vxll9\") pod \"redhat-operators-2skv5\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.192008 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-catalog-content\") pod \"redhat-operators-2skv5\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.192353 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-catalog-content\") pod \"redhat-operators-2skv5\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.192585 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-utilities\") pod \"redhat-operators-2skv5\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.217023 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxll9\" (UniqueName: \"kubernetes.io/projected/af6b10ff-6922-44fc-a489-514cbcffe69d-kube-api-access-vxll9\") pod \"redhat-operators-2skv5\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.269423 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" event={"ID":"f2261944-710e-493f-9360-de4dea650ae5","Type":"ContainerStarted","Data":"b263e364ded51d0ec1708e37a2240252abf3690d0519b50caac9fc1eb1b58bdb"} Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.269493 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" event={"ID":"f2261944-710e-493f-9360-de4dea650ae5","Type":"ContainerStarted","Data":"9a868f8cda24244d43f0996bd5b9cd61871f1c34b168efa5688d7ed8ff520fb9"} Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.269511 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.272638 4667 generic.go:334] "Generic (PLEG): container finished" podID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerID="ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98" exitCode=0 Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.272700 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62g78" event={"ID":"cb52efdd-e2d7-493e-b786-34eb5e260d40","Type":"ContainerDied","Data":"ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98"} Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.272722 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62g78" event={"ID":"cb52efdd-e2d7-493e-b786-34eb5e260d40","Type":"ContainerStarted","Data":"af36384da35e347f41f935c8fdd6f4151388f0f96f4de798a6440b35b1ab538d"} Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.274591 4667 generic.go:334] "Generic (PLEG): container finished" podID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerID="fad31b024f0b60bd70635c958fbde94db0e6efbcaee02b3b0d8b1c3c3d0e8e6c" exitCode=0 Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.274619 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w8pd" event={"ID":"3acdff5b-96b2-4100-a595-3c17295f9ef6","Type":"ContainerDied","Data":"fad31b024f0b60bd70635c958fbde94db0e6efbcaee02b3b0d8b1c3c3d0e8e6c"} Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.274654 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w8pd" event={"ID":"3acdff5b-96b2-4100-a595-3c17295f9ef6","Type":"ContainerStarted","Data":"399a0425620f55ed940d7f309ebab2766c5ba5bc2b2a4b1e4dff06cb61635b24"} Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.285300 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" podStartSLOduration=125.285287334 podStartE2EDuration="2m5.285287334s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:55.281051886 +0000 UTC m=+143.778898655" watchObservedRunningTime="2025-09-29 17:11:55.285287334 +0000 UTC m=+143.783134102" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.291162 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.296370 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1831c9f5-ebda-412a-bd70-eb0176387e9b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1831c9f5-ebda-412a-bd70-eb0176387e9b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.296477 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1831c9f5-ebda-412a-bd70-eb0176387e9b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1831c9f5-ebda-412a-bd70-eb0176387e9b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.351492 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:55 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:55 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:55 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.351537 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.381128 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zr8bb"] Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.384330 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.385168 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zr8bb"] Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.401569 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1831c9f5-ebda-412a-bd70-eb0176387e9b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1831c9f5-ebda-412a-bd70-eb0176387e9b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.410159 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1831c9f5-ebda-412a-bd70-eb0176387e9b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1831c9f5-ebda-412a-bd70-eb0176387e9b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.402316 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1831c9f5-ebda-412a-bd70-eb0176387e9b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"1831c9f5-ebda-412a-bd70-eb0176387e9b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.432415 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1831c9f5-ebda-412a-bd70-eb0176387e9b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"1831c9f5-ebda-412a-bd70-eb0176387e9b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.479336 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.486065 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.514372 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-utilities\") pod \"redhat-operators-zr8bb\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.514487 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wwt6\" (UniqueName: \"kubernetes.io/projected/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-kube-api-access-4wwt6\") pod \"redhat-operators-zr8bb\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.514553 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-catalog-content\") pod \"redhat-operators-zr8bb\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.598221 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2skv5"] Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.618361 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-config-volume\") pod \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.618416 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47rrb\" (UniqueName: \"kubernetes.io/projected/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-kube-api-access-47rrb\") pod \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.618471 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-secret-volume\") pod \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\" (UID: \"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2\") " Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.618670 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-catalog-content\") pod \"redhat-operators-zr8bb\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.618764 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-utilities\") pod \"redhat-operators-zr8bb\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.618829 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wwt6\" (UniqueName: \"kubernetes.io/projected/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-kube-api-access-4wwt6\") pod \"redhat-operators-zr8bb\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.619193 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-config-volume" (OuterVolumeSpecName: "config-volume") pod "68e21fe0-a72c-42e2-b2a8-89f8eed86bd2" (UID: "68e21fe0-a72c-42e2-b2a8-89f8eed86bd2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.619364 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-catalog-content\") pod \"redhat-operators-zr8bb\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.619547 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-utilities\") pod \"redhat-operators-zr8bb\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.622964 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "68e21fe0-a72c-42e2-b2a8-89f8eed86bd2" (UID: "68e21fe0-a72c-42e2-b2a8-89f8eed86bd2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.623826 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-kube-api-access-47rrb" (OuterVolumeSpecName: "kube-api-access-47rrb") pod "68e21fe0-a72c-42e2-b2a8-89f8eed86bd2" (UID: "68e21fe0-a72c-42e2-b2a8-89f8eed86bd2"). InnerVolumeSpecName "kube-api-access-47rrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.639261 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wwt6\" (UniqueName: \"kubernetes.io/projected/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-kube-api-access-4wwt6\") pod \"redhat-operators-zr8bb\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.715131 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.720005 4667 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.720029 4667 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.720040 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47rrb\" (UniqueName: \"kubernetes.io/projected/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2-kube-api-access-47rrb\") on node \"crc\" DevicePath \"\"" Sep 29 17:11:55 crc kubenswrapper[4667]: W0929 17:11:55.740859 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod1831c9f5_ebda_412a_bd70_eb0176387e9b.slice/crio-3406572b12d2668c671be6c8f13ef981230007cd5300f1be247d87be79b9407d WatchSource:0}: Error finding container 3406572b12d2668c671be6c8f13ef981230007cd5300f1be247d87be79b9407d: Status 404 returned error can't find the container with id 3406572b12d2668c671be6c8f13ef981230007cd5300f1be247d87be79b9407d Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.749878 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.833601 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Sep 29 17:11:55 crc kubenswrapper[4667]: I0929 17:11:55.938914 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zr8bb"] Sep 29 17:11:55 crc kubenswrapper[4667]: W0929 17:11:55.959173 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cb7d916_bbc3_4ab7_8d3d_258e70bcd5b0.slice/crio-3aff62edac91cd4f71e27a8efb0d6e2afff4e7cebcf41149c1bd85062f137cee WatchSource:0}: Error finding container 3aff62edac91cd4f71e27a8efb0d6e2afff4e7cebcf41149c1bd85062f137cee: Status 404 returned error can't find the container with id 3aff62edac91cd4f71e27a8efb0d6e2afff4e7cebcf41149c1bd85062f137cee Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.279303 4667 generic.go:334] "Generic (PLEG): container finished" podID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerID="e5d9fb7843dd098db77ca72171608747e4421707f590860b5aa2f036021c3fff" exitCode=0 Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.279354 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2skv5" event={"ID":"af6b10ff-6922-44fc-a489-514cbcffe69d","Type":"ContainerDied","Data":"e5d9fb7843dd098db77ca72171608747e4421707f590860b5aa2f036021c3fff"} Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.279378 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2skv5" event={"ID":"af6b10ff-6922-44fc-a489-514cbcffe69d","Type":"ContainerStarted","Data":"ff174d5d51bd327d536909e91b42159b13867293756d0d9b1c798e0ea28b145c"} Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.281623 4667 generic.go:334] "Generic (PLEG): container finished" podID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerID="46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83" exitCode=0 Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.281967 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zr8bb" event={"ID":"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0","Type":"ContainerDied","Data":"46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83"} Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.282980 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zr8bb" event={"ID":"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0","Type":"ContainerStarted","Data":"3aff62edac91cd4f71e27a8efb0d6e2afff4e7cebcf41149c1bd85062f137cee"} Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.285046 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" event={"ID":"68e21fe0-a72c-42e2-b2a8-89f8eed86bd2","Type":"ContainerDied","Data":"08b4de814a7d084b04508e672c7e361b202fd222b3411a5241a96459484d2c6c"} Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.285065 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08b4de814a7d084b04508e672c7e361b202fd222b3411a5241a96459484d2c6c" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.285104 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.288177 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1831c9f5-ebda-412a-bd70-eb0176387e9b","Type":"ContainerStarted","Data":"31cf5c62726bf89547ac145c13c0f9f82b84e963573d02a9952cc199b7acfb43"} Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.288204 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1831c9f5-ebda-412a-bd70-eb0176387e9b","Type":"ContainerStarted","Data":"3406572b12d2668c671be6c8f13ef981230007cd5300f1be247d87be79b9407d"} Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.323747 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.323729934 podStartE2EDuration="1.323729934s" podCreationTimestamp="2025-09-29 17:11:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:11:56.310600072 +0000 UTC m=+144.808446841" watchObservedRunningTime="2025-09-29 17:11:56.323729934 +0000 UTC m=+144.821576703" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.350082 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:56 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:56 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:56 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.350134 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.583176 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 29 17:11:56 crc kubenswrapper[4667]: E0929 17:11:56.583596 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68e21fe0-a72c-42e2-b2a8-89f8eed86bd2" containerName="collect-profiles" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.583608 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="68e21fe0-a72c-42e2-b2a8-89f8eed86bd2" containerName="collect-profiles" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.583695 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="68e21fe0-a72c-42e2-b2a8-89f8eed86bd2" containerName="collect-profiles" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.584036 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.586214 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.591146 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.592227 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.734923 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/543853b4-7225-442a-89b9-91ecf1bde696-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"543853b4-7225-442a-89b9-91ecf1bde696\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.735011 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/543853b4-7225-442a-89b9-91ecf1bde696-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"543853b4-7225-442a-89b9-91ecf1bde696\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.836107 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/543853b4-7225-442a-89b9-91ecf1bde696-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"543853b4-7225-442a-89b9-91ecf1bde696\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.836175 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/543853b4-7225-442a-89b9-91ecf1bde696-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"543853b4-7225-442a-89b9-91ecf1bde696\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.836257 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/543853b4-7225-442a-89b9-91ecf1bde696-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"543853b4-7225-442a-89b9-91ecf1bde696\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.852270 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/543853b4-7225-442a-89b9-91ecf1bde696-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"543853b4-7225-442a-89b9-91ecf1bde696\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.897562 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.897620 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:11:56 crc kubenswrapper[4667]: I0929 17:11:56.904409 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.295562 4667 generic.go:334] "Generic (PLEG): container finished" podID="1831c9f5-ebda-412a-bd70-eb0176387e9b" containerID="31cf5c62726bf89547ac145c13c0f9f82b84e963573d02a9952cc199b7acfb43" exitCode=0 Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.295599 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1831c9f5-ebda-412a-bd70-eb0176387e9b","Type":"ContainerDied","Data":"31cf5c62726bf89547ac145c13c0f9f82b84e963573d02a9952cc199b7acfb43"} Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.350425 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:57 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:57 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:57 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.350470 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.745515 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.745631 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.745701 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.745789 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.746591 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.748919 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.749646 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.749763 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.925165 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.936186 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.939166 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-b89jg" Sep 29 17:11:57 crc kubenswrapper[4667]: I0929 17:11:57.939729 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Sep 29 17:11:58 crc kubenswrapper[4667]: I0929 17:11:58.037098 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:58 crc kubenswrapper[4667]: I0929 17:11:58.047085 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rj4qf" Sep 29 17:11:58 crc kubenswrapper[4667]: I0929 17:11:58.219261 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:58 crc kubenswrapper[4667]: I0929 17:11:58.219303 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:11:58 crc kubenswrapper[4667]: I0929 17:11:58.221323 4667 patch_prober.go:28] interesting pod/console-f9d7485db-qhc55 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Sep 29 17:11:58 crc kubenswrapper[4667]: I0929 17:11:58.221371 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-qhc55" podUID="cab548cd-ee10-421c-9648-02dd9ad58dfa" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Sep 29 17:11:58 crc kubenswrapper[4667]: I0929 17:11:58.348297 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:11:58 crc kubenswrapper[4667]: I0929 17:11:58.352668 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:58 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:58 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:58 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:58 crc kubenswrapper[4667]: I0929 17:11:58.352707 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.349265 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:11:59 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:11:59 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:11:59 crc kubenswrapper[4667]: healthz check failed Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.349343 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.520823 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.571332 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1831c9f5-ebda-412a-bd70-eb0176387e9b-kube-api-access\") pod \"1831c9f5-ebda-412a-bd70-eb0176387e9b\" (UID: \"1831c9f5-ebda-412a-bd70-eb0176387e9b\") " Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.572092 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1831c9f5-ebda-412a-bd70-eb0176387e9b-kubelet-dir\") pod \"1831c9f5-ebda-412a-bd70-eb0176387e9b\" (UID: \"1831c9f5-ebda-412a-bd70-eb0176387e9b\") " Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.572240 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1831c9f5-ebda-412a-bd70-eb0176387e9b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1831c9f5-ebda-412a-bd70-eb0176387e9b" (UID: "1831c9f5-ebda-412a-bd70-eb0176387e9b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.572423 4667 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1831c9f5-ebda-412a-bd70-eb0176387e9b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.576200 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1831c9f5-ebda-412a-bd70-eb0176387e9b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1831c9f5-ebda-412a-bd70-eb0176387e9b" (UID: "1831c9f5-ebda-412a-bd70-eb0176387e9b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.673910 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1831c9f5-ebda-412a-bd70-eb0176387e9b-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 17:11:59 crc kubenswrapper[4667]: I0929 17:11:59.757481 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Sep 29 17:11:59 crc kubenswrapper[4667]: W0929 17:11:59.815864 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-e9ca8062675799efe064e94bbf373c301d9db70abf841de4b8ba9cda8264a97b WatchSource:0}: Error finding container e9ca8062675799efe064e94bbf373c301d9db70abf841de4b8ba9cda8264a97b: Status 404 returned error can't find the container with id e9ca8062675799efe064e94bbf373c301d9db70abf841de4b8ba9cda8264a97b Sep 29 17:11:59 crc kubenswrapper[4667]: W0929 17:11:59.923191 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-3786bd7e8bb23c906ce88a854aa4479915d2592a0d48f22bdfd2c8e2193e94d9 WatchSource:0}: Error finding container 3786bd7e8bb23c906ce88a854aa4479915d2592a0d48f22bdfd2c8e2193e94d9: Status 404 returned error can't find the container with id 3786bd7e8bb23c906ce88a854aa4479915d2592a0d48f22bdfd2c8e2193e94d9 Sep 29 17:11:59 crc kubenswrapper[4667]: W0929 17:11:59.925688 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-5bb34b8a990e7fe8a199363587e1f97249a44fb2d7f56307dfe51395a0c12078 WatchSource:0}: Error finding container 5bb34b8a990e7fe8a199363587e1f97249a44fb2d7f56307dfe51395a0c12078: Status 404 returned error can't find the container with id 5bb34b8a990e7fe8a199363587e1f97249a44fb2d7f56307dfe51395a0c12078 Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.354540 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.354633 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"1831c9f5-ebda-412a-bd70-eb0176387e9b","Type":"ContainerDied","Data":"3406572b12d2668c671be6c8f13ef981230007cd5300f1be247d87be79b9407d"} Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.356903 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3406572b12d2668c671be6c8f13ef981230007cd5300f1be247d87be79b9407d" Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.360086 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:12:00 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:12:00 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:12:00 crc kubenswrapper[4667]: healthz check failed Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.360126 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.366016 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"543853b4-7225-442a-89b9-91ecf1bde696","Type":"ContainerStarted","Data":"2d9e37bf4d386707913be19af073e03a117bdfdc92d47f695b6f183b3241aaae"} Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.366083 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"543853b4-7225-442a-89b9-91ecf1bde696","Type":"ContainerStarted","Data":"405e300eaf7f5db7b44b12559ecd8e9e9fd4275040391f96dd3506cd8d353c42"} Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.371035 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"160e6a44a8fbe616207cb42350a7148fab3fb9f63954a4dda9a87af76ebf6bac"} Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.371067 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e9ca8062675799efe064e94bbf373c301d9db70abf841de4b8ba9cda8264a97b"} Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.373707 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"27207ffc9aedf5ba1c3824d4f716354bc74b98a80e58ed7beb5dca84404cfbc9"} Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.373734 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5bb34b8a990e7fe8a199363587e1f97249a44fb2d7f56307dfe51395a0c12078"} Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.375996 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c7632d7d97bb468e6503ee4fdfc8f150d0bf7419ca9d15481cd0297d451c89c9"} Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.376059 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3786bd7e8bb23c906ce88a854aa4479915d2592a0d48f22bdfd2c8e2193e94d9"} Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.376227 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:12:00 crc kubenswrapper[4667]: I0929 17:12:00.386613 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.386597161 podStartE2EDuration="4.386597161s" podCreationTimestamp="2025-09-29 17:11:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:12:00.382433608 +0000 UTC m=+148.880280377" watchObservedRunningTime="2025-09-29 17:12:00.386597161 +0000 UTC m=+148.884443930" Sep 29 17:12:01 crc kubenswrapper[4667]: I0929 17:12:01.350370 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:12:01 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:12:01 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:12:01 crc kubenswrapper[4667]: healthz check failed Sep 29 17:12:01 crc kubenswrapper[4667]: I0929 17:12:01.350429 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:12:01 crc kubenswrapper[4667]: I0929 17:12:01.410515 4667 generic.go:334] "Generic (PLEG): container finished" podID="543853b4-7225-442a-89b9-91ecf1bde696" containerID="2d9e37bf4d386707913be19af073e03a117bdfdc92d47f695b6f183b3241aaae" exitCode=0 Sep 29 17:12:01 crc kubenswrapper[4667]: I0929 17:12:01.411350 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"543853b4-7225-442a-89b9-91ecf1bde696","Type":"ContainerDied","Data":"2d9e37bf4d386707913be19af073e03a117bdfdc92d47f695b6f183b3241aaae"} Sep 29 17:12:02 crc kubenswrapper[4667]: I0929 17:12:02.350823 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:12:02 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:12:02 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:12:02 crc kubenswrapper[4667]: healthz check failed Sep 29 17:12:02 crc kubenswrapper[4667]: I0929 17:12:02.351117 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:12:02 crc kubenswrapper[4667]: I0929 17:12:02.643939 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:12:02 crc kubenswrapper[4667]: I0929 17:12:02.726289 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/543853b4-7225-442a-89b9-91ecf1bde696-kube-api-access\") pod \"543853b4-7225-442a-89b9-91ecf1bde696\" (UID: \"543853b4-7225-442a-89b9-91ecf1bde696\") " Sep 29 17:12:02 crc kubenswrapper[4667]: I0929 17:12:02.726370 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/543853b4-7225-442a-89b9-91ecf1bde696-kubelet-dir\") pod \"543853b4-7225-442a-89b9-91ecf1bde696\" (UID: \"543853b4-7225-442a-89b9-91ecf1bde696\") " Sep 29 17:12:02 crc kubenswrapper[4667]: I0929 17:12:02.726575 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/543853b4-7225-442a-89b9-91ecf1bde696-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "543853b4-7225-442a-89b9-91ecf1bde696" (UID: "543853b4-7225-442a-89b9-91ecf1bde696"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:12:02 crc kubenswrapper[4667]: I0929 17:12:02.731774 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/543853b4-7225-442a-89b9-91ecf1bde696-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "543853b4-7225-442a-89b9-91ecf1bde696" (UID: "543853b4-7225-442a-89b9-91ecf1bde696"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:12:02 crc kubenswrapper[4667]: I0929 17:12:02.827910 4667 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/543853b4-7225-442a-89b9-91ecf1bde696-kubelet-dir\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:02 crc kubenswrapper[4667]: I0929 17:12:02.828226 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/543853b4-7225-442a-89b9-91ecf1bde696-kube-api-access\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:03 crc kubenswrapper[4667]: I0929 17:12:03.350995 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:12:03 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:12:03 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:12:03 crc kubenswrapper[4667]: healthz check failed Sep 29 17:12:03 crc kubenswrapper[4667]: I0929 17:12:03.351094 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:12:03 crc kubenswrapper[4667]: I0929 17:12:03.441363 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"543853b4-7225-442a-89b9-91ecf1bde696","Type":"ContainerDied","Data":"405e300eaf7f5db7b44b12559ecd8e9e9fd4275040391f96dd3506cd8d353c42"} Sep 29 17:12:03 crc kubenswrapper[4667]: I0929 17:12:03.441415 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="405e300eaf7f5db7b44b12559ecd8e9e9fd4275040391f96dd3506cd8d353c42" Sep 29 17:12:03 crc kubenswrapper[4667]: I0929 17:12:03.441515 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Sep 29 17:12:03 crc kubenswrapper[4667]: I0929 17:12:03.796350 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-xbcfz" Sep 29 17:12:04 crc kubenswrapper[4667]: I0929 17:12:04.350739 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:12:04 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:12:04 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:12:04 crc kubenswrapper[4667]: healthz check failed Sep 29 17:12:04 crc kubenswrapper[4667]: I0929 17:12:04.350816 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:12:05 crc kubenswrapper[4667]: I0929 17:12:05.349472 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:12:05 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:12:05 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:12:05 crc kubenswrapper[4667]: healthz check failed Sep 29 17:12:05 crc kubenswrapper[4667]: I0929 17:12:05.349727 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:12:06 crc kubenswrapper[4667]: I0929 17:12:06.349568 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:12:06 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:12:06 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:12:06 crc kubenswrapper[4667]: healthz check failed Sep 29 17:12:06 crc kubenswrapper[4667]: I0929 17:12:06.349626 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:12:07 crc kubenswrapper[4667]: I0929 17:12:07.350661 4667 patch_prober.go:28] interesting pod/router-default-5444994796-lwt9b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Sep 29 17:12:07 crc kubenswrapper[4667]: [-]has-synced failed: reason withheld Sep 29 17:12:07 crc kubenswrapper[4667]: [+]process-running ok Sep 29 17:12:07 crc kubenswrapper[4667]: healthz check failed Sep 29 17:12:07 crc kubenswrapper[4667]: I0929 17:12:07.350717 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lwt9b" podUID="eacfedb4-33c9-4744-9399-cf1848bb0353" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Sep 29 17:12:08 crc kubenswrapper[4667]: I0929 17:12:08.219418 4667 patch_prober.go:28] interesting pod/console-f9d7485db-qhc55 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Sep 29 17:12:08 crc kubenswrapper[4667]: I0929 17:12:08.219472 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-qhc55" podUID="cab548cd-ee10-421c-9648-02dd9ad58dfa" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Sep 29 17:12:08 crc kubenswrapper[4667]: I0929 17:12:08.350050 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:12:08 crc kubenswrapper[4667]: I0929 17:12:08.352088 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lwt9b" Sep 29 17:12:11 crc kubenswrapper[4667]: I0929 17:12:11.646223 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:12:11 crc kubenswrapper[4667]: I0929 17:12:11.652368 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d360e6c4-2b40-4214-bb7c-5d08038c1b62-metrics-certs\") pod \"network-metrics-daemon-cl5p9\" (UID: \"d360e6c4-2b40-4214-bb7c-5d08038c1b62\") " pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:12:11 crc kubenswrapper[4667]: I0929 17:12:11.730129 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-cl5p9" Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.305008 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-cl5p9"] Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.484624 4667 generic.go:334] "Generic (PLEG): container finished" podID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerID="c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b" exitCode=0 Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.484700 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62g78" event={"ID":"cb52efdd-e2d7-493e-b786-34eb5e260d40","Type":"ContainerDied","Data":"c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b"} Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.486519 4667 generic.go:334] "Generic (PLEG): container finished" podID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerID="c34c22e95616f9c2ce252a9b85f28f986dd7dad21ad511d055e7f06add83fd83" exitCode=0 Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.486596 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n26jh" event={"ID":"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b","Type":"ContainerDied","Data":"c34c22e95616f9c2ce252a9b85f28f986dd7dad21ad511d055e7f06add83fd83"} Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.491397 4667 generic.go:334] "Generic (PLEG): container finished" podID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerID="5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18" exitCode=0 Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.491467 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5t5jw" event={"ID":"dcbcc4f2-3dff-49dc-b310-8504021366d3","Type":"ContainerDied","Data":"5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18"} Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.492726 4667 generic.go:334] "Generic (PLEG): container finished" podID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerID="3f834fad05ce89206f93b49c0d65891cb1d0a93697e8b35f08c255307026992f" exitCode=0 Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.492795 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w8pd" event={"ID":"3acdff5b-96b2-4100-a595-3c17295f9ef6","Type":"ContainerDied","Data":"3f834fad05ce89206f93b49c0d65891cb1d0a93697e8b35f08c255307026992f"} Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.494804 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2skv5" event={"ID":"af6b10ff-6922-44fc-a489-514cbcffe69d","Type":"ContainerStarted","Data":"0716c93cea1cad2038490500775e3a93588b5b491e2e5d66cb90187cec6006a9"} Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.496400 4667 generic.go:334] "Generic (PLEG): container finished" podID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerID="badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4" exitCode=0 Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.496431 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zr8bb" event={"ID":"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0","Type":"ContainerDied","Data":"badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4"} Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.498701 4667 generic.go:334] "Generic (PLEG): container finished" podID="28489e28-fd28-4526-99dd-4f72aefbc272" containerID="b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4" exitCode=0 Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.498763 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2s5vw" event={"ID":"28489e28-fd28-4526-99dd-4f72aefbc272","Type":"ContainerDied","Data":"b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4"} Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.500828 4667 generic.go:334] "Generic (PLEG): container finished" podID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerID="880a0bac667d6aed8e825340494bc387104a515d72ad0666735e0553de4c9a17" exitCode=0 Sep 29 17:12:12 crc kubenswrapper[4667]: I0929 17:12:12.500880 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7rjk" event={"ID":"3795f5d0-58d5-40a1-bfcb-98da8a26e905","Type":"ContainerDied","Data":"880a0bac667d6aed8e825340494bc387104a515d72ad0666735e0553de4c9a17"} Sep 29 17:12:12 crc kubenswrapper[4667]: W0929 17:12:12.510735 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd360e6c4_2b40_4214_bb7c_5d08038c1b62.slice/crio-1314f2f0f3d58217cb898a1c610a3108b4853b3169ffb18fb54dcdc3faf2942e WatchSource:0}: Error finding container 1314f2f0f3d58217cb898a1c610a3108b4853b3169ffb18fb54dcdc3faf2942e: Status 404 returned error can't find the container with id 1314f2f0f3d58217cb898a1c610a3108b4853b3169ffb18fb54dcdc3faf2942e Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.506708 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62g78" event={"ID":"cb52efdd-e2d7-493e-b786-34eb5e260d40","Type":"ContainerStarted","Data":"3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.508449 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w8pd" event={"ID":"3acdff5b-96b2-4100-a595-3c17295f9ef6","Type":"ContainerStarted","Data":"119be153c5b3e0dd61c32bb894839ee41f3270a3ba217544ad10a5e64e65034d"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.510886 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7rjk" event={"ID":"3795f5d0-58d5-40a1-bfcb-98da8a26e905","Type":"ContainerStarted","Data":"58c455b551a0cf0801671ae5f347abebcf24559362a6c0e14766a5cc2145bbd2"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.513562 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zr8bb" event={"ID":"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0","Type":"ContainerStarted","Data":"da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.515392 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2s5vw" event={"ID":"28489e28-fd28-4526-99dd-4f72aefbc272","Type":"ContainerStarted","Data":"8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.516963 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" event={"ID":"d360e6c4-2b40-4214-bb7c-5d08038c1b62","Type":"ContainerStarted","Data":"fa25361f26318644196752e29f1d7dd28b6790ed2e07cdcf1c00fb0e056f4271"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.517004 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" event={"ID":"d360e6c4-2b40-4214-bb7c-5d08038c1b62","Type":"ContainerStarted","Data":"dd8ff8cca2683adc379985975d1be56c8e54e3b6fe38cf43fe85cd15a7dd1673"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.517015 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-cl5p9" event={"ID":"d360e6c4-2b40-4214-bb7c-5d08038c1b62","Type":"ContainerStarted","Data":"1314f2f0f3d58217cb898a1c610a3108b4853b3169ffb18fb54dcdc3faf2942e"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.518799 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n26jh" event={"ID":"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b","Type":"ContainerStarted","Data":"b0b0e73b415f6e42889ee1e787f542e230cf95e5d1097dada75ffb332e916dfe"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.520927 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5t5jw" event={"ID":"dcbcc4f2-3dff-49dc-b310-8504021366d3","Type":"ContainerStarted","Data":"08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.522449 4667 generic.go:334] "Generic (PLEG): container finished" podID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerID="0716c93cea1cad2038490500775e3a93588b5b491e2e5d66cb90187cec6006a9" exitCode=0 Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.522478 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2skv5" event={"ID":"af6b10ff-6922-44fc-a489-514cbcffe69d","Type":"ContainerDied","Data":"0716c93cea1cad2038490500775e3a93588b5b491e2e5d66cb90187cec6006a9"} Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.527401 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-62g78" podStartSLOduration=1.831082766 podStartE2EDuration="19.527393179s" podCreationTimestamp="2025-09-29 17:11:54 +0000 UTC" firstStartedPulling="2025-09-29 17:11:55.273734694 +0000 UTC m=+143.771581462" lastFinishedPulling="2025-09-29 17:12:12.970045105 +0000 UTC m=+161.467891875" observedRunningTime="2025-09-29 17:12:13.524127097 +0000 UTC m=+162.021973866" watchObservedRunningTime="2025-09-29 17:12:13.527393179 +0000 UTC m=+162.025239948" Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.560061 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n26jh" podStartSLOduration=2.723393727 podStartE2EDuration="22.560049348s" podCreationTimestamp="2025-09-29 17:11:51 +0000 UTC" firstStartedPulling="2025-09-29 17:11:53.242086542 +0000 UTC m=+141.739933310" lastFinishedPulling="2025-09-29 17:12:13.078742162 +0000 UTC m=+161.576588931" observedRunningTime="2025-09-29 17:12:13.546099012 +0000 UTC m=+162.043945781" watchObservedRunningTime="2025-09-29 17:12:13.560049348 +0000 UTC m=+162.057896118" Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.573916 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zr8bb" podStartSLOduration=1.7208810570000002 podStartE2EDuration="18.573904174s" podCreationTimestamp="2025-09-29 17:11:55 +0000 UTC" firstStartedPulling="2025-09-29 17:11:56.28560868 +0000 UTC m=+144.783455449" lastFinishedPulling="2025-09-29 17:12:13.138631797 +0000 UTC m=+161.636478566" observedRunningTime="2025-09-29 17:12:13.572057838 +0000 UTC m=+162.069904606" watchObservedRunningTime="2025-09-29 17:12:13.573904174 +0000 UTC m=+162.071750943" Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.590378 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-cl5p9" podStartSLOduration=143.590369909 podStartE2EDuration="2m23.590369909s" podCreationTimestamp="2025-09-29 17:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:12:13.586561616 +0000 UTC m=+162.084408385" watchObservedRunningTime="2025-09-29 17:12:13.590369909 +0000 UTC m=+162.088216678" Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.601253 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z7rjk" podStartSLOduration=2.651944309 podStartE2EDuration="22.601246175s" podCreationTimestamp="2025-09-29 17:11:51 +0000 UTC" firstStartedPulling="2025-09-29 17:11:53.232329954 +0000 UTC m=+141.730176723" lastFinishedPulling="2025-09-29 17:12:13.18163182 +0000 UTC m=+161.679478589" observedRunningTime="2025-09-29 17:12:13.601040256 +0000 UTC m=+162.098887026" watchObservedRunningTime="2025-09-29 17:12:13.601246175 +0000 UTC m=+162.099092944" Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.636707 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5t5jw" podStartSLOduration=2.756087089 podStartE2EDuration="21.636699884s" podCreationTimestamp="2025-09-29 17:11:52 +0000 UTC" firstStartedPulling="2025-09-29 17:11:54.257479326 +0000 UTC m=+142.755326094" lastFinishedPulling="2025-09-29 17:12:13.138092121 +0000 UTC m=+161.635938889" observedRunningTime="2025-09-29 17:12:13.619893638 +0000 UTC m=+162.117740408" watchObservedRunningTime="2025-09-29 17:12:13.636699884 +0000 UTC m=+162.134546653" Sep 29 17:12:13 crc kubenswrapper[4667]: I0929 17:12:13.651616 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2s5vw" podStartSLOduration=2.912339678 podStartE2EDuration="21.651609496s" podCreationTimestamp="2025-09-29 17:11:52 +0000 UTC" firstStartedPulling="2025-09-29 17:11:54.267451939 +0000 UTC m=+142.765298708" lastFinishedPulling="2025-09-29 17:12:13.006721758 +0000 UTC m=+161.504568526" observedRunningTime="2025-09-29 17:12:13.637282641 +0000 UTC m=+162.135129410" watchObservedRunningTime="2025-09-29 17:12:13.651609496 +0000 UTC m=+162.149456266" Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.289684 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.290504 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.290543 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.305530 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5w8pd" podStartSLOduration=3.612502512 podStartE2EDuration="21.305507607s" podCreationTimestamp="2025-09-29 17:11:53 +0000 UTC" firstStartedPulling="2025-09-29 17:11:55.279383303 +0000 UTC m=+143.777230073" lastFinishedPulling="2025-09-29 17:12:12.972388399 +0000 UTC m=+161.470235168" observedRunningTime="2025-09-29 17:12:13.653633659 +0000 UTC m=+162.151480428" watchObservedRunningTime="2025-09-29 17:12:14.305507607 +0000 UTC m=+162.803354375" Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.387393 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.529756 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2skv5" event={"ID":"af6b10ff-6922-44fc-a489-514cbcffe69d","Type":"ContainerStarted","Data":"b17798d5c520003906a98ca2c51d824423547775f0b37616a920a10c56fa89cd"} Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.556755 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2skv5" podStartSLOduration=2.842654826 podStartE2EDuration="20.556741599s" podCreationTimestamp="2025-09-29 17:11:54 +0000 UTC" firstStartedPulling="2025-09-29 17:11:56.285094792 +0000 UTC m=+144.782941562" lastFinishedPulling="2025-09-29 17:12:13.999181566 +0000 UTC m=+162.497028335" observedRunningTime="2025-09-29 17:12:14.554255336 +0000 UTC m=+163.052102105" watchObservedRunningTime="2025-09-29 17:12:14.556741599 +0000 UTC m=+163.054588368" Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.707415 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.707478 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:12:14 crc kubenswrapper[4667]: I0929 17:12:14.741526 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:12:15 crc kubenswrapper[4667]: I0929 17:12:15.291724 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:12:15 crc kubenswrapper[4667]: I0929 17:12:15.291978 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:12:15 crc kubenswrapper[4667]: I0929 17:12:15.750210 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:12:15 crc kubenswrapper[4667]: I0929 17:12:15.750378 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:12:16 crc kubenswrapper[4667]: I0929 17:12:16.317746 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2skv5" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerName="registry-server" probeResult="failure" output=< Sep 29 17:12:16 crc kubenswrapper[4667]: timeout: failed to connect service ":50051" within 1s Sep 29 17:12:16 crc kubenswrapper[4667]: > Sep 29 17:12:16 crc kubenswrapper[4667]: I0929 17:12:16.777458 4667 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zr8bb" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerName="registry-server" probeResult="failure" output=< Sep 29 17:12:16 crc kubenswrapper[4667]: timeout: failed to connect service ":50051" within 1s Sep 29 17:12:16 crc kubenswrapper[4667]: > Sep 29 17:12:18 crc kubenswrapper[4667]: I0929 17:12:18.223411 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:12:18 crc kubenswrapper[4667]: I0929 17:12:18.226449 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.130238 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.130532 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.158160 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.301919 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.302286 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.329837 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.504422 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.504624 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.532409 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.585245 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.585827 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.587719 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.894138 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.894175 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:12:22 crc kubenswrapper[4667]: I0929 17:12:22.924473 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:12:23 crc kubenswrapper[4667]: I0929 17:12:23.589215 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:12:24 crc kubenswrapper[4667]: I0929 17:12:24.178875 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2s5vw"] Sep 29 17:12:24 crc kubenswrapper[4667]: I0929 17:12:24.319644 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:12:24 crc kubenswrapper[4667]: I0929 17:12:24.734886 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:12:24 crc kubenswrapper[4667]: I0929 17:12:24.776971 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5t5jw"] Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.320087 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.345755 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.573720 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5t5jw" podUID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerName="registry-server" containerID="cri-o://08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210" gracePeriod=2 Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.574033 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2s5vw" podUID="28489e28-fd28-4526-99dd-4f72aefbc272" containerName="registry-server" containerID="cri-o://8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60" gracePeriod=2 Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.777784 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.820276 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.865056 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.908189 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.998068 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-utilities\") pod \"dcbcc4f2-3dff-49dc-b310-8504021366d3\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.998127 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-utilities\") pod \"28489e28-fd28-4526-99dd-4f72aefbc272\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.998155 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-catalog-content\") pod \"dcbcc4f2-3dff-49dc-b310-8504021366d3\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.998218 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-catalog-content\") pod \"28489e28-fd28-4526-99dd-4f72aefbc272\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.998272 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvt94\" (UniqueName: \"kubernetes.io/projected/28489e28-fd28-4526-99dd-4f72aefbc272-kube-api-access-fvt94\") pod \"28489e28-fd28-4526-99dd-4f72aefbc272\" (UID: \"28489e28-fd28-4526-99dd-4f72aefbc272\") " Sep 29 17:12:25 crc kubenswrapper[4667]: I0929 17:12:25.998327 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn92v\" (UniqueName: \"kubernetes.io/projected/dcbcc4f2-3dff-49dc-b310-8504021366d3-kube-api-access-qn92v\") pod \"dcbcc4f2-3dff-49dc-b310-8504021366d3\" (UID: \"dcbcc4f2-3dff-49dc-b310-8504021366d3\") " Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:25.998671 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-utilities" (OuterVolumeSpecName: "utilities") pod "28489e28-fd28-4526-99dd-4f72aefbc272" (UID: "28489e28-fd28-4526-99dd-4f72aefbc272"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:25.998684 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-utilities" (OuterVolumeSpecName: "utilities") pod "dcbcc4f2-3dff-49dc-b310-8504021366d3" (UID: "dcbcc4f2-3dff-49dc-b310-8504021366d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.003543 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcbcc4f2-3dff-49dc-b310-8504021366d3-kube-api-access-qn92v" (OuterVolumeSpecName: "kube-api-access-qn92v") pod "dcbcc4f2-3dff-49dc-b310-8504021366d3" (UID: "dcbcc4f2-3dff-49dc-b310-8504021366d3"). InnerVolumeSpecName "kube-api-access-qn92v". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.003576 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28489e28-fd28-4526-99dd-4f72aefbc272-kube-api-access-fvt94" (OuterVolumeSpecName: "kube-api-access-fvt94") pod "28489e28-fd28-4526-99dd-4f72aefbc272" (UID: "28489e28-fd28-4526-99dd-4f72aefbc272"). InnerVolumeSpecName "kube-api-access-fvt94". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.029832 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dcbcc4f2-3dff-49dc-b310-8504021366d3" (UID: "dcbcc4f2-3dff-49dc-b310-8504021366d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.036738 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28489e28-fd28-4526-99dd-4f72aefbc272" (UID: "28489e28-fd28-4526-99dd-4f72aefbc272"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.099079 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvt94\" (UniqueName: \"kubernetes.io/projected/28489e28-fd28-4526-99dd-4f72aefbc272-kube-api-access-fvt94\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.099103 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn92v\" (UniqueName: \"kubernetes.io/projected/dcbcc4f2-3dff-49dc-b310-8504021366d3-kube-api-access-qn92v\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.099113 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.099123 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.099131 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dcbcc4f2-3dff-49dc-b310-8504021366d3-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.099138 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28489e28-fd28-4526-99dd-4f72aefbc272-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.579161 4667 generic.go:334] "Generic (PLEG): container finished" podID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerID="08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210" exitCode=0 Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.579231 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5t5jw" event={"ID":"dcbcc4f2-3dff-49dc-b310-8504021366d3","Type":"ContainerDied","Data":"08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210"} Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.579242 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5t5jw" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.579256 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5t5jw" event={"ID":"dcbcc4f2-3dff-49dc-b310-8504021366d3","Type":"ContainerDied","Data":"0f2373ee6e55b86dee82dd899d123f32c95ebeec0c18e6f5cb61f5a9d49b3d41"} Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.579272 4667 scope.go:117] "RemoveContainer" containerID="08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.581495 4667 generic.go:334] "Generic (PLEG): container finished" podID="28489e28-fd28-4526-99dd-4f72aefbc272" containerID="8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60" exitCode=0 Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.582045 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2s5vw" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.586401 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2s5vw" event={"ID":"28489e28-fd28-4526-99dd-4f72aefbc272","Type":"ContainerDied","Data":"8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60"} Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.586428 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2s5vw" event={"ID":"28489e28-fd28-4526-99dd-4f72aefbc272","Type":"ContainerDied","Data":"5840676216b8c371277806c29734544fefb27a895baddeeedeb4d425e8ff46fc"} Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.594679 4667 scope.go:117] "RemoveContainer" containerID="5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.608525 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z6xxn"] Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.611872 4667 scope.go:117] "RemoveContainer" containerID="21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.616044 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5t5jw"] Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.620831 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5t5jw"] Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.633296 4667 scope.go:117] "RemoveContainer" containerID="08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210" Sep 29 17:12:26 crc kubenswrapper[4667]: E0929 17:12:26.633823 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210\": container with ID starting with 08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210 not found: ID does not exist" containerID="08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.633891 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210"} err="failed to get container status \"08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210\": rpc error: code = NotFound desc = could not find container \"08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210\": container with ID starting with 08f0b57e2b03138f6a2e583267426c7de3ef268e4ec1ba1887496c42c83b0210 not found: ID does not exist" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.633925 4667 scope.go:117] "RemoveContainer" containerID="5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18" Sep 29 17:12:26 crc kubenswrapper[4667]: E0929 17:12:26.634192 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18\": container with ID starting with 5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18 not found: ID does not exist" containerID="5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.634212 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18"} err="failed to get container status \"5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18\": rpc error: code = NotFound desc = could not find container \"5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18\": container with ID starting with 5fe2d8b3759f48892bff86f4ad0e2698dfef3d5afe3a75de20df3b2c4bcb6e18 not found: ID does not exist" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.634226 4667 scope.go:117] "RemoveContainer" containerID="21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd" Sep 29 17:12:26 crc kubenswrapper[4667]: E0929 17:12:26.634428 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd\": container with ID starting with 21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd not found: ID does not exist" containerID="21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.634448 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd"} err="failed to get container status \"21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd\": rpc error: code = NotFound desc = could not find container \"21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd\": container with ID starting with 21977b20bfbe71dc54bfd8a00718832f85c6bca73cd4a9d3bbcffea979cdaafd not found: ID does not exist" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.634460 4667 scope.go:117] "RemoveContainer" containerID="8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.640050 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2s5vw"] Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.643895 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2s5vw"] Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.647208 4667 scope.go:117] "RemoveContainer" containerID="b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.667981 4667 scope.go:117] "RemoveContainer" containerID="6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.681965 4667 scope.go:117] "RemoveContainer" containerID="8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60" Sep 29 17:12:26 crc kubenswrapper[4667]: E0929 17:12:26.682278 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60\": container with ID starting with 8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60 not found: ID does not exist" containerID="8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.682316 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60"} err="failed to get container status \"8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60\": rpc error: code = NotFound desc = could not find container \"8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60\": container with ID starting with 8dfe4f938bd7ff7fb701bce905fbcf71610553b3a44e424373352709e3a6df60 not found: ID does not exist" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.682348 4667 scope.go:117] "RemoveContainer" containerID="b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4" Sep 29 17:12:26 crc kubenswrapper[4667]: E0929 17:12:26.682612 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4\": container with ID starting with b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4 not found: ID does not exist" containerID="b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.682652 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4"} err="failed to get container status \"b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4\": rpc error: code = NotFound desc = could not find container \"b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4\": container with ID starting with b9439bb4d9214e64a2495d3acbfc914051ff1c265d170288670a42ecf8efe8f4 not found: ID does not exist" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.682679 4667 scope.go:117] "RemoveContainer" containerID="6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285" Sep 29 17:12:26 crc kubenswrapper[4667]: E0929 17:12:26.682967 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285\": container with ID starting with 6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285 not found: ID does not exist" containerID="6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.682994 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285"} err="failed to get container status \"6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285\": rpc error: code = NotFound desc = could not find container \"6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285\": container with ID starting with 6d079ab8a50e9c2a47a60bbb28b7ab1051b45c09b95b8588e9e359d0794a3285 not found: ID does not exist" Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.897370 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:12:26 crc kubenswrapper[4667]: I0929 17:12:26.897420 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.178763 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-62g78"] Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.178960 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-62g78" podUID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerName="registry-server" containerID="cri-o://3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed" gracePeriod=2 Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.463267 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.522910 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-catalog-content\") pod \"cb52efdd-e2d7-493e-b786-34eb5e260d40\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.522990 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-utilities\") pod \"cb52efdd-e2d7-493e-b786-34eb5e260d40\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.523021 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgxxk\" (UniqueName: \"kubernetes.io/projected/cb52efdd-e2d7-493e-b786-34eb5e260d40-kube-api-access-hgxxk\") pod \"cb52efdd-e2d7-493e-b786-34eb5e260d40\" (UID: \"cb52efdd-e2d7-493e-b786-34eb5e260d40\") " Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.523596 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-utilities" (OuterVolumeSpecName: "utilities") pod "cb52efdd-e2d7-493e-b786-34eb5e260d40" (UID: "cb52efdd-e2d7-493e-b786-34eb5e260d40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.529968 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb52efdd-e2d7-493e-b786-34eb5e260d40-kube-api-access-hgxxk" (OuterVolumeSpecName: "kube-api-access-hgxxk") pod "cb52efdd-e2d7-493e-b786-34eb5e260d40" (UID: "cb52efdd-e2d7-493e-b786-34eb5e260d40"). InnerVolumeSpecName "kube-api-access-hgxxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.532637 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb52efdd-e2d7-493e-b786-34eb5e260d40" (UID: "cb52efdd-e2d7-493e-b786-34eb5e260d40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.586801 4667 generic.go:334] "Generic (PLEG): container finished" podID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerID="3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed" exitCode=0 Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.586881 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62g78" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.586877 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62g78" event={"ID":"cb52efdd-e2d7-493e-b786-34eb5e260d40","Type":"ContainerDied","Data":"3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed"} Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.587047 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62g78" event={"ID":"cb52efdd-e2d7-493e-b786-34eb5e260d40","Type":"ContainerDied","Data":"af36384da35e347f41f935c8fdd6f4151388f0f96f4de798a6440b35b1ab538d"} Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.587084 4667 scope.go:117] "RemoveContainer" containerID="3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.599041 4667 scope.go:117] "RemoveContainer" containerID="c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.612747 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-62g78"] Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.613011 4667 scope.go:117] "RemoveContainer" containerID="ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.614512 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-62g78"] Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.624575 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.624593 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb52efdd-e2d7-493e-b786-34eb5e260d40-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.624603 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgxxk\" (UniqueName: \"kubernetes.io/projected/cb52efdd-e2d7-493e-b786-34eb5e260d40-kube-api-access-hgxxk\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.635806 4667 scope.go:117] "RemoveContainer" containerID="3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed" Sep 29 17:12:27 crc kubenswrapper[4667]: E0929 17:12:27.636139 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed\": container with ID starting with 3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed not found: ID does not exist" containerID="3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.636162 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed"} err="failed to get container status \"3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed\": rpc error: code = NotFound desc = could not find container \"3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed\": container with ID starting with 3fec89b99d1b5cca998c585b1f2a5f924ecd88d8462640af707e1a278c5cf9ed not found: ID does not exist" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.636180 4667 scope.go:117] "RemoveContainer" containerID="c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b" Sep 29 17:12:27 crc kubenswrapper[4667]: E0929 17:12:27.636385 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b\": container with ID starting with c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b not found: ID does not exist" containerID="c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.636403 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b"} err="failed to get container status \"c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b\": rpc error: code = NotFound desc = could not find container \"c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b\": container with ID starting with c0b16e2f1b71579aa1c23a0445d293e097a3a3d2321ca0e41855cd1436fc657b not found: ID does not exist" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.636415 4667 scope.go:117] "RemoveContainer" containerID="ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98" Sep 29 17:12:27 crc kubenswrapper[4667]: E0929 17:12:27.636643 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98\": container with ID starting with ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98 not found: ID does not exist" containerID="ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.636668 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98"} err="failed to get container status \"ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98\": rpc error: code = NotFound desc = could not find container \"ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98\": container with ID starting with ab9508db59048a175bf8d9bd343c49221ca71cebf8c656503bd2e5def5f9af98 not found: ID does not exist" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.821824 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28489e28-fd28-4526-99dd-4f72aefbc272" path="/var/lib/kubelet/pods/28489e28-fd28-4526-99dd-4f72aefbc272/volumes" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.822387 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb52efdd-e2d7-493e-b786-34eb5e260d40" path="/var/lib/kubelet/pods/cb52efdd-e2d7-493e-b786-34eb5e260d40/volumes" Sep 29 17:12:27 crc kubenswrapper[4667]: I0929 17:12:27.822889 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcbcc4f2-3dff-49dc-b310-8504021366d3" path="/var/lib/kubelet/pods/dcbcc4f2-3dff-49dc-b310-8504021366d3/volumes" Sep 29 17:12:28 crc kubenswrapper[4667]: I0929 17:12:28.745894 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mfkxj" Sep 29 17:12:29 crc kubenswrapper[4667]: I0929 17:12:29.577480 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zr8bb"] Sep 29 17:12:29 crc kubenswrapper[4667]: I0929 17:12:29.577646 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zr8bb" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerName="registry-server" containerID="cri-o://da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c" gracePeriod=2 Sep 29 17:12:29 crc kubenswrapper[4667]: I0929 17:12:29.848802 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:12:29 crc kubenswrapper[4667]: I0929 17:12:29.946244 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wwt6\" (UniqueName: \"kubernetes.io/projected/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-kube-api-access-4wwt6\") pod \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " Sep 29 17:12:29 crc kubenswrapper[4667]: I0929 17:12:29.946322 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-utilities\") pod \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " Sep 29 17:12:29 crc kubenswrapper[4667]: I0929 17:12:29.946387 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-catalog-content\") pod \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\" (UID: \"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0\") " Sep 29 17:12:29 crc kubenswrapper[4667]: I0929 17:12:29.947004 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-utilities" (OuterVolumeSpecName: "utilities") pod "0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" (UID: "0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:12:29 crc kubenswrapper[4667]: I0929 17:12:29.950255 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-kube-api-access-4wwt6" (OuterVolumeSpecName: "kube-api-access-4wwt6") pod "0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" (UID: "0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0"). InnerVolumeSpecName "kube-api-access-4wwt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.003014 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" (UID: "0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.047773 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.047801 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wwt6\" (UniqueName: \"kubernetes.io/projected/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-kube-api-access-4wwt6\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.047815 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.601620 4667 generic.go:334] "Generic (PLEG): container finished" podID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerID="da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c" exitCode=0 Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.601662 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zr8bb" event={"ID":"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0","Type":"ContainerDied","Data":"da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c"} Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.601687 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zr8bb" event={"ID":"0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0","Type":"ContainerDied","Data":"3aff62edac91cd4f71e27a8efb0d6e2afff4e7cebcf41149c1bd85062f137cee"} Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.601704 4667 scope.go:117] "RemoveContainer" containerID="da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.601708 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zr8bb" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.615964 4667 scope.go:117] "RemoveContainer" containerID="badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.625162 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zr8bb"] Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.627095 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zr8bb"] Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.651039 4667 scope.go:117] "RemoveContainer" containerID="46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.661500 4667 scope.go:117] "RemoveContainer" containerID="da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c" Sep 29 17:12:30 crc kubenswrapper[4667]: E0929 17:12:30.661975 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c\": container with ID starting with da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c not found: ID does not exist" containerID="da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.662007 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c"} err="failed to get container status \"da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c\": rpc error: code = NotFound desc = could not find container \"da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c\": container with ID starting with da416e01a25a12a30a244e18a27b3fd6315bc52218ddbfd49c16bb20aab9108c not found: ID does not exist" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.662028 4667 scope.go:117] "RemoveContainer" containerID="badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4" Sep 29 17:12:30 crc kubenswrapper[4667]: E0929 17:12:30.662248 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4\": container with ID starting with badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4 not found: ID does not exist" containerID="badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.662267 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4"} err="failed to get container status \"badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4\": rpc error: code = NotFound desc = could not find container \"badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4\": container with ID starting with badad453fdc283ed8e4d161d75cffbc57f3c4a5dbe881f754c16a0ef68917da4 not found: ID does not exist" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.662279 4667 scope.go:117] "RemoveContainer" containerID="46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83" Sep 29 17:12:30 crc kubenswrapper[4667]: E0929 17:12:30.662490 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83\": container with ID starting with 46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83 not found: ID does not exist" containerID="46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83" Sep 29 17:12:30 crc kubenswrapper[4667]: I0929 17:12:30.662511 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83"} err="failed to get container status \"46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83\": rpc error: code = NotFound desc = could not find container \"46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83\": container with ID starting with 46836723ae8c37c47fd06524c0e1618a3c09444a4f875d7b4b3eff3fd8606b83 not found: ID does not exist" Sep 29 17:12:31 crc kubenswrapper[4667]: I0929 17:12:31.827137 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" path="/var/lib/kubelet/pods/0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0/volumes" Sep 29 17:12:37 crc kubenswrapper[4667]: I0929 17:12:37.940562 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.628219 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" podUID="f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" containerName="oauth-openshift" containerID="cri-o://ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a" gracePeriod=15 Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.918899 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.956694 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-c8559d799-c4jz7"] Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962656 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28489e28-fd28-4526-99dd-4f72aefbc272" containerName="extract-utilities" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962689 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="28489e28-fd28-4526-99dd-4f72aefbc272" containerName="extract-utilities" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962702 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="543853b4-7225-442a-89b9-91ecf1bde696" containerName="pruner" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962708 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="543853b4-7225-442a-89b9-91ecf1bde696" containerName="pruner" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962716 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerName="extract-utilities" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962723 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerName="extract-utilities" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962730 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerName="extract-utilities" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962736 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerName="extract-utilities" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962743 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962749 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962762 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerName="extract-utilities" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962768 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerName="extract-utilities" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962777 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerName="extract-content" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962783 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerName="extract-content" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962791 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerName="extract-content" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962796 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerName="extract-content" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962806 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" containerName="oauth-openshift" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962811 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" containerName="oauth-openshift" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962817 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962822 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962830 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1831c9f5-ebda-412a-bd70-eb0176387e9b" containerName="pruner" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962835 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="1831c9f5-ebda-412a-bd70-eb0176387e9b" containerName="pruner" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962855 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28489e28-fd28-4526-99dd-4f72aefbc272" containerName="extract-content" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962861 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="28489e28-fd28-4526-99dd-4f72aefbc272" containerName="extract-content" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962868 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962875 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962887 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerName="extract-content" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962893 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerName="extract-content" Sep 29 17:12:51 crc kubenswrapper[4667]: E0929 17:12:51.962901 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28489e28-fd28-4526-99dd-4f72aefbc272" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.962907 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="28489e28-fd28-4526-99dd-4f72aefbc272" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.963041 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="543853b4-7225-442a-89b9-91ecf1bde696" containerName="pruner" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.963050 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb7d916-bbc3-4ab7-8d3d-258e70bcd5b0" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.963058 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb52efdd-e2d7-493e-b786-34eb5e260d40" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.963071 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="28489e28-fd28-4526-99dd-4f72aefbc272" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.963078 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcbcc4f2-3dff-49dc-b310-8504021366d3" containerName="registry-server" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.963084 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" containerName="oauth-openshift" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.963090 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="1831c9f5-ebda-412a-bd70-eb0176387e9b" containerName="pruner" Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.963459 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-c8559d799-c4jz7"] Sep 29 17:12:51 crc kubenswrapper[4667]: I0929 17:12:51.963550 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084047 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-dir\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084273 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7nbt\" (UniqueName: \"kubernetes.io/projected/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-kube-api-access-z7nbt\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084417 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-cliconfig\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084495 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-ocp-branding-template\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084561 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-session\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084644 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-login\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084188 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084787 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-trusted-ca-bundle\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084885 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-serving-cert\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.084984 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-idp-0-file-data\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.085061 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-policies\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.085141 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-router-certs\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.085252 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-service-ca\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.085826 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-provider-selection\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.085962 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-error\") pod \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\" (UID: \"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8\") " Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.085775 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.085884 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.085964 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.085987 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.086558 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.086703 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-router-certs\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.086872 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-service-ca\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.087029 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.087142 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a550572a-40e4-488b-862e-c626d775f03f-audit-dir\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.087246 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.087434 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-template-error\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.087565 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.087696 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-session\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.087819 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.088087 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkqbf\" (UniqueName: \"kubernetes.io/projected/a550572a-40e4-488b-862e-c626d775f03f-kube-api-access-tkqbf\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.088212 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-audit-policies\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.088335 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-template-login\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.088468 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.088599 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.088668 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.088728 4667 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-policies\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.088781 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.088837 4667 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-audit-dir\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.090554 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.090771 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.090979 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.091084 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-kube-api-access-z7nbt" (OuterVolumeSpecName: "kube-api-access-z7nbt") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "kube-api-access-z7nbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.091146 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.091643 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.091836 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.091999 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.092175 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" (UID: "f92e52ac-cc8e-4af2-8ec2-2d81800f17d8"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190124 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-audit-policies\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190164 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-template-login\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190192 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190223 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190238 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-router-certs\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190262 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-service-ca\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190288 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190308 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a550572a-40e4-488b-862e-c626d775f03f-audit-dir\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190325 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190342 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-template-error\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190364 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190385 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-session\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190401 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190416 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkqbf\" (UniqueName: \"kubernetes.io/projected/a550572a-40e4-488b-862e-c626d775f03f-kube-api-access-tkqbf\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190453 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190465 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190475 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190486 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190495 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190504 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7nbt\" (UniqueName: \"kubernetes.io/projected/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-kube-api-access-z7nbt\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190513 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190521 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190530 4667 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190922 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a550572a-40e4-488b-862e-c626d775f03f-audit-dir\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.190993 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-audit-policies\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.191550 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.192101 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-service-ca\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.193037 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.193889 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.193910 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-template-login\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.194079 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.194358 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.194802 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-router-certs\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.195051 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.195488 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-system-session\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.197494 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a550572a-40e4-488b-862e-c626d775f03f-v4-0-config-user-template-error\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.205750 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkqbf\" (UniqueName: \"kubernetes.io/projected/a550572a-40e4-488b-862e-c626d775f03f-kube-api-access-tkqbf\") pod \"oauth-openshift-c8559d799-c4jz7\" (UID: \"a550572a-40e4-488b-862e-c626d775f03f\") " pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.282907 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.446504 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-c8559d799-c4jz7"] Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.686580 4667 generic.go:334] "Generic (PLEG): container finished" podID="f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" containerID="ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a" exitCode=0 Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.686653 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.686688 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" event={"ID":"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8","Type":"ContainerDied","Data":"ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a"} Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.687016 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-z6xxn" event={"ID":"f92e52ac-cc8e-4af2-8ec2-2d81800f17d8","Type":"ContainerDied","Data":"77ea1044b8b4ea03f52c14faa214e849993a8ab5b8dc0d82b029eaf66f3f8b99"} Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.687045 4667 scope.go:117] "RemoveContainer" containerID="ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.688044 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" event={"ID":"a550572a-40e4-488b-862e-c626d775f03f","Type":"ContainerStarted","Data":"0b2c680b2e0df9415198bce2b1924f6a2918172499fb3c6cf246f528935c9e5e"} Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.688081 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" event={"ID":"a550572a-40e4-488b-862e-c626d775f03f","Type":"ContainerStarted","Data":"779af0bfa3d0f276c91ac290b63f3b5ab38dbc3e17ccb0ffde4d88f7b9a20ece"} Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.688292 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.689530 4667 patch_prober.go:28] interesting pod/oauth-openshift-c8559d799-c4jz7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.54:6443/healthz\": dial tcp 10.217.0.54:6443: connect: connection refused" start-of-body= Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.689567 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" podUID="a550572a-40e4-488b-862e-c626d775f03f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.54:6443/healthz\": dial tcp 10.217.0.54:6443: connect: connection refused" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.699787 4667 scope.go:117] "RemoveContainer" containerID="ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a" Sep 29 17:12:52 crc kubenswrapper[4667]: E0929 17:12:52.700081 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a\": container with ID starting with ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a not found: ID does not exist" containerID="ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.700114 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a"} err="failed to get container status \"ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a\": rpc error: code = NotFound desc = could not find container \"ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a\": container with ID starting with ce0e40362fb5858bf9bc3ac391aba8f21fc086b85c517ccf48220a0cb6d1fa9a not found: ID does not exist" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.709128 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" podStartSLOduration=26.709112255 podStartE2EDuration="26.709112255s" podCreationTimestamp="2025-09-29 17:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:12:52.707467667 +0000 UTC m=+201.205314436" watchObservedRunningTime="2025-09-29 17:12:52.709112255 +0000 UTC m=+201.206959023" Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.717306 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z6xxn"] Sep 29 17:12:52 crc kubenswrapper[4667]: I0929 17:12:52.719653 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-z6xxn"] Sep 29 17:12:53 crc kubenswrapper[4667]: I0929 17:12:53.698219 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-c8559d799-c4jz7" Sep 29 17:12:53 crc kubenswrapper[4667]: I0929 17:12:53.826399 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f92e52ac-cc8e-4af2-8ec2-2d81800f17d8" path="/var/lib/kubelet/pods/f92e52ac-cc8e-4af2-8ec2-2d81800f17d8/volumes" Sep 29 17:12:56 crc kubenswrapper[4667]: I0929 17:12:56.897132 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:12:56 crc kubenswrapper[4667]: I0929 17:12:56.897352 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:12:56 crc kubenswrapper[4667]: I0929 17:12:56.897389 4667 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:12:56 crc kubenswrapper[4667]: I0929 17:12:56.897740 4667 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657"} pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 17:12:56 crc kubenswrapper[4667]: I0929 17:12:56.897784 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" containerID="cri-o://2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657" gracePeriod=600 Sep 29 17:12:57 crc kubenswrapper[4667]: I0929 17:12:57.714264 4667 generic.go:334] "Generic (PLEG): container finished" podID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerID="2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657" exitCode=0 Sep 29 17:12:57 crc kubenswrapper[4667]: I0929 17:12:57.714341 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerDied","Data":"2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657"} Sep 29 17:12:57 crc kubenswrapper[4667]: I0929 17:12:57.714571 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"a9ba7db32c2d86b3e6b9098efebccb5be88d2dd5c0dcb0c69d78ec2f5c309f2f"} Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.540268 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z7rjk"] Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.541250 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z7rjk" podUID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerName="registry-server" containerID="cri-o://58c455b551a0cf0801671ae5f347abebcf24559362a6c0e14766a5cc2145bbd2" gracePeriod=30 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.554688 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n26jh"] Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.554935 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n26jh" podUID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerName="registry-server" containerID="cri-o://b0b0e73b415f6e42889ee1e787f542e230cf95e5d1097dada75ffb332e916dfe" gracePeriod=30 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.560177 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wjdtk"] Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.560363 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" podUID="56908f94-88ac-4298-98e5-3ba1315bd4ae" containerName="marketplace-operator" containerID="cri-o://d6efad009de94276a9d0733d1f15f31a8bf612324fb895731346587f8e2d3fad" gracePeriod=30 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.568913 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w8pd"] Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.569083 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5w8pd" podUID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerName="registry-server" containerID="cri-o://119be153c5b3e0dd61c32bb894839ee41f3270a3ba217544ad10a5e64e65034d" gracePeriod=30 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.573938 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qzqrj"] Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.574501 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.579316 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2skv5"] Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.579452 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2skv5" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerName="registry-server" containerID="cri-o://b17798d5c520003906a98ca2c51d824423547775f0b37616a920a10c56fa89cd" gracePeriod=30 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.582709 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qzqrj"] Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.685728 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8a1767f-4534-4fa7-b126-e9c0ba822350-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qzqrj\" (UID: \"f8a1767f-4534-4fa7-b126-e9c0ba822350\") " pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.685903 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57pfq\" (UniqueName: \"kubernetes.io/projected/f8a1767f-4534-4fa7-b126-e9c0ba822350-kube-api-access-57pfq\") pod \"marketplace-operator-79b997595-qzqrj\" (UID: \"f8a1767f-4534-4fa7-b126-e9c0ba822350\") " pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.686026 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8a1767f-4534-4fa7-b126-e9c0ba822350-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qzqrj\" (UID: \"f8a1767f-4534-4fa7-b126-e9c0ba822350\") " pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.767025 4667 generic.go:334] "Generic (PLEG): container finished" podID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerID="119be153c5b3e0dd61c32bb894839ee41f3270a3ba217544ad10a5e64e65034d" exitCode=0 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.767199 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w8pd" event={"ID":"3acdff5b-96b2-4100-a595-3c17295f9ef6","Type":"ContainerDied","Data":"119be153c5b3e0dd61c32bb894839ee41f3270a3ba217544ad10a5e64e65034d"} Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.768926 4667 generic.go:334] "Generic (PLEG): container finished" podID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerID="b17798d5c520003906a98ca2c51d824423547775f0b37616a920a10c56fa89cd" exitCode=0 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.768968 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2skv5" event={"ID":"af6b10ff-6922-44fc-a489-514cbcffe69d","Type":"ContainerDied","Data":"b17798d5c520003906a98ca2c51d824423547775f0b37616a920a10c56fa89cd"} Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.770604 4667 generic.go:334] "Generic (PLEG): container finished" podID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerID="58c455b551a0cf0801671ae5f347abebcf24559362a6c0e14766a5cc2145bbd2" exitCode=0 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.770641 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7rjk" event={"ID":"3795f5d0-58d5-40a1-bfcb-98da8a26e905","Type":"ContainerDied","Data":"58c455b551a0cf0801671ae5f347abebcf24559362a6c0e14766a5cc2145bbd2"} Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.771508 4667 generic.go:334] "Generic (PLEG): container finished" podID="56908f94-88ac-4298-98e5-3ba1315bd4ae" containerID="d6efad009de94276a9d0733d1f15f31a8bf612324fb895731346587f8e2d3fad" exitCode=0 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.771541 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" event={"ID":"56908f94-88ac-4298-98e5-3ba1315bd4ae","Type":"ContainerDied","Data":"d6efad009de94276a9d0733d1f15f31a8bf612324fb895731346587f8e2d3fad"} Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.773126 4667 generic.go:334] "Generic (PLEG): container finished" podID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerID="b0b0e73b415f6e42889ee1e787f542e230cf95e5d1097dada75ffb332e916dfe" exitCode=0 Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.773155 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n26jh" event={"ID":"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b","Type":"ContainerDied","Data":"b0b0e73b415f6e42889ee1e787f542e230cf95e5d1097dada75ffb332e916dfe"} Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.787392 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8a1767f-4534-4fa7-b126-e9c0ba822350-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qzqrj\" (UID: \"f8a1767f-4534-4fa7-b126-e9c0ba822350\") " pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.787428 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57pfq\" (UniqueName: \"kubernetes.io/projected/f8a1767f-4534-4fa7-b126-e9c0ba822350-kube-api-access-57pfq\") pod \"marketplace-operator-79b997595-qzqrj\" (UID: \"f8a1767f-4534-4fa7-b126-e9c0ba822350\") " pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.787468 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8a1767f-4534-4fa7-b126-e9c0ba822350-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qzqrj\" (UID: \"f8a1767f-4534-4fa7-b126-e9c0ba822350\") " pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.788522 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8a1767f-4534-4fa7-b126-e9c0ba822350-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qzqrj\" (UID: \"f8a1767f-4534-4fa7-b126-e9c0ba822350\") " pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.792779 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f8a1767f-4534-4fa7-b126-e9c0ba822350-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qzqrj\" (UID: \"f8a1767f-4534-4fa7-b126-e9c0ba822350\") " pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.802434 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57pfq\" (UniqueName: \"kubernetes.io/projected/f8a1767f-4534-4fa7-b126-e9c0ba822350-kube-api-access-57pfq\") pod \"marketplace-operator-79b997595-qzqrj\" (UID: \"f8a1767f-4534-4fa7-b126-e9c0ba822350\") " pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.968030 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.971232 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.975332 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.978996 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.985535 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:13:10 crc kubenswrapper[4667]: I0929 17:13:10.996720 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091387 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w67n5\" (UniqueName: \"kubernetes.io/projected/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-kube-api-access-w67n5\") pod \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091431 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-catalog-content\") pod \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091453 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-utilities\") pod \"af6b10ff-6922-44fc-a489-514cbcffe69d\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091529 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-catalog-content\") pod \"af6b10ff-6922-44fc-a489-514cbcffe69d\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091581 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-utilities\") pod \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091596 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-catalog-content\") pod \"3acdff5b-96b2-4100-a595-3c17295f9ef6\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091623 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-utilities\") pod \"3acdff5b-96b2-4100-a595-3c17295f9ef6\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091646 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-utilities\") pod \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091672 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxll9\" (UniqueName: \"kubernetes.io/projected/af6b10ff-6922-44fc-a489-514cbcffe69d-kube-api-access-vxll9\") pod \"af6b10ff-6922-44fc-a489-514cbcffe69d\" (UID: \"af6b10ff-6922-44fc-a489-514cbcffe69d\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091690 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chwt7\" (UniqueName: \"kubernetes.io/projected/3795f5d0-58d5-40a1-bfcb-98da8a26e905-kube-api-access-chwt7\") pod \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\" (UID: \"3795f5d0-58d5-40a1-bfcb-98da8a26e905\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091714 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-catalog-content\") pod \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\" (UID: \"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.091728 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjxhv\" (UniqueName: \"kubernetes.io/projected/3acdff5b-96b2-4100-a595-3c17295f9ef6-kube-api-access-rjxhv\") pod \"3acdff5b-96b2-4100-a595-3c17295f9ef6\" (UID: \"3acdff5b-96b2-4100-a595-3c17295f9ef6\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.093384 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-utilities" (OuterVolumeSpecName: "utilities") pod "eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" (UID: "eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.094074 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-utilities" (OuterVolumeSpecName: "utilities") pod "af6b10ff-6922-44fc-a489-514cbcffe69d" (UID: "af6b10ff-6922-44fc-a489-514cbcffe69d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.094104 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-utilities" (OuterVolumeSpecName: "utilities") pod "3acdff5b-96b2-4100-a595-3c17295f9ef6" (UID: "3acdff5b-96b2-4100-a595-3c17295f9ef6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.095059 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-utilities" (OuterVolumeSpecName: "utilities") pod "3795f5d0-58d5-40a1-bfcb-98da8a26e905" (UID: "3795f5d0-58d5-40a1-bfcb-98da8a26e905"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.095859 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3795f5d0-58d5-40a1-bfcb-98da8a26e905-kube-api-access-chwt7" (OuterVolumeSpecName: "kube-api-access-chwt7") pod "3795f5d0-58d5-40a1-bfcb-98da8a26e905" (UID: "3795f5d0-58d5-40a1-bfcb-98da8a26e905"). InnerVolumeSpecName "kube-api-access-chwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.097572 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af6b10ff-6922-44fc-a489-514cbcffe69d-kube-api-access-vxll9" (OuterVolumeSpecName: "kube-api-access-vxll9") pod "af6b10ff-6922-44fc-a489-514cbcffe69d" (UID: "af6b10ff-6922-44fc-a489-514cbcffe69d"). InnerVolumeSpecName "kube-api-access-vxll9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.104760 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3acdff5b-96b2-4100-a595-3c17295f9ef6" (UID: "3acdff5b-96b2-4100-a595-3c17295f9ef6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.105189 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-kube-api-access-w67n5" (OuterVolumeSpecName: "kube-api-access-w67n5") pod "eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" (UID: "eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b"). InnerVolumeSpecName "kube-api-access-w67n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.110363 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3acdff5b-96b2-4100-a595-3c17295f9ef6-kube-api-access-rjxhv" (OuterVolumeSpecName: "kube-api-access-rjxhv") pod "3acdff5b-96b2-4100-a595-3c17295f9ef6" (UID: "3acdff5b-96b2-4100-a595-3c17295f9ef6"). InnerVolumeSpecName "kube-api-access-rjxhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.132232 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3795f5d0-58d5-40a1-bfcb-98da8a26e905" (UID: "3795f5d0-58d5-40a1-bfcb-98da8a26e905"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.137611 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" (UID: "eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.156424 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af6b10ff-6922-44fc-a489-514cbcffe69d" (UID: "af6b10ff-6922-44fc-a489-514cbcffe69d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193172 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtvd2\" (UniqueName: \"kubernetes.io/projected/56908f94-88ac-4298-98e5-3ba1315bd4ae-kube-api-access-vtvd2\") pod \"56908f94-88ac-4298-98e5-3ba1315bd4ae\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193259 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-trusted-ca\") pod \"56908f94-88ac-4298-98e5-3ba1315bd4ae\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193367 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-operator-metrics\") pod \"56908f94-88ac-4298-98e5-3ba1315bd4ae\" (UID: \"56908f94-88ac-4298-98e5-3ba1315bd4ae\") " Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193566 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w67n5\" (UniqueName: \"kubernetes.io/projected/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-kube-api-access-w67n5\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193663 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193679 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193740 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af6b10ff-6922-44fc-a489-514cbcffe69d-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193751 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193759 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193767 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3acdff5b-96b2-4100-a595-3c17295f9ef6-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193774 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3795f5d0-58d5-40a1-bfcb-98da8a26e905-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193781 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxll9\" (UniqueName: \"kubernetes.io/projected/af6b10ff-6922-44fc-a489-514cbcffe69d-kube-api-access-vxll9\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193789 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chwt7\" (UniqueName: \"kubernetes.io/projected/3795f5d0-58d5-40a1-bfcb-98da8a26e905-kube-api-access-chwt7\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193796 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.193804 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjxhv\" (UniqueName: \"kubernetes.io/projected/3acdff5b-96b2-4100-a595-3c17295f9ef6-kube-api-access-rjxhv\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.194082 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "56908f94-88ac-4298-98e5-3ba1315bd4ae" (UID: "56908f94-88ac-4298-98e5-3ba1315bd4ae"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.196129 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56908f94-88ac-4298-98e5-3ba1315bd4ae-kube-api-access-vtvd2" (OuterVolumeSpecName: "kube-api-access-vtvd2") pod "56908f94-88ac-4298-98e5-3ba1315bd4ae" (UID: "56908f94-88ac-4298-98e5-3ba1315bd4ae"). InnerVolumeSpecName "kube-api-access-vtvd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.196365 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "56908f94-88ac-4298-98e5-3ba1315bd4ae" (UID: "56908f94-88ac-4298-98e5-3ba1315bd4ae"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.294833 4667 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.294870 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtvd2\" (UniqueName: \"kubernetes.io/projected/56908f94-88ac-4298-98e5-3ba1315bd4ae-kube-api-access-vtvd2\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.294880 4667 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56908f94-88ac-4298-98e5-3ba1315bd4ae-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.338006 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qzqrj"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.780145 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5w8pd" event={"ID":"3acdff5b-96b2-4100-a595-3c17295f9ef6","Type":"ContainerDied","Data":"399a0425620f55ed940d7f309ebab2766c5ba5bc2b2a4b1e4dff06cb61635b24"} Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.780187 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5w8pd" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.780519 4667 scope.go:117] "RemoveContainer" containerID="119be153c5b3e0dd61c32bb894839ee41f3270a3ba217544ad10a5e64e65034d" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.783173 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2skv5" event={"ID":"af6b10ff-6922-44fc-a489-514cbcffe69d","Type":"ContainerDied","Data":"ff174d5d51bd327d536909e91b42159b13867293756d0d9b1c798e0ea28b145c"} Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.783191 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2skv5" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.785161 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7rjk" event={"ID":"3795f5d0-58d5-40a1-bfcb-98da8a26e905","Type":"ContainerDied","Data":"813b59d342db02717ed58b05a4e6f01b8b30d556e3c8eda487878fa9eecec532"} Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.785297 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7rjk" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.788673 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" event={"ID":"56908f94-88ac-4298-98e5-3ba1315bd4ae","Type":"ContainerDied","Data":"175d85e12165607e664750dc01b891ce89fde0e0fcd07ea3ccceac90424bb52f"} Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.788735 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wjdtk" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.793241 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" event={"ID":"f8a1767f-4534-4fa7-b126-e9c0ba822350","Type":"ContainerStarted","Data":"5dfa5ef2c03e1df8d8a2bae6aa04ddb70350a6f5954beef01d7eba0c0d3f55e6"} Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.793303 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.793315 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" event={"ID":"f8a1767f-4534-4fa7-b126-e9c0ba822350","Type":"ContainerStarted","Data":"757f205470d25fda3730f51f3a6617d0555a9f68ad9add777843036f2396682d"} Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.793591 4667 scope.go:117] "RemoveContainer" containerID="3f834fad05ce89206f93b49c0d65891cb1d0a93697e8b35f08c255307026992f" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.796622 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.796647 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n26jh" event={"ID":"eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b","Type":"ContainerDied","Data":"0fe4580fb39db149701f2684e926f1c8e9889df65d6bf0fb42b51c245504300d"} Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.796659 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n26jh" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.806101 4667 scope.go:117] "RemoveContainer" containerID="fad31b024f0b60bd70635c958fbde94db0e6efbcaee02b3b0d8b1c3c3d0e8e6c" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.808069 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qzqrj" podStartSLOduration=1.8080573869999998 podStartE2EDuration="1.808057387s" podCreationTimestamp="2025-09-29 17:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:13:11.807013432 +0000 UTC m=+220.304860201" watchObservedRunningTime="2025-09-29 17:13:11.808057387 +0000 UTC m=+220.305904187" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.821942 4667 scope.go:117] "RemoveContainer" containerID="b17798d5c520003906a98ca2c51d824423547775f0b37616a920a10c56fa89cd" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.824695 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w8pd"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.824728 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5w8pd"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.833598 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2skv5"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.835472 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2skv5"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.842981 4667 scope.go:117] "RemoveContainer" containerID="0716c93cea1cad2038490500775e3a93588b5b491e2e5d66cb90187cec6006a9" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.855520 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n26jh"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.858708 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n26jh"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.864657 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wjdtk"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.867289 4667 scope.go:117] "RemoveContainer" containerID="e5d9fb7843dd098db77ca72171608747e4421707f590860b5aa2f036021c3fff" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.873176 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wjdtk"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.884249 4667 scope.go:117] "RemoveContainer" containerID="58c455b551a0cf0801671ae5f347abebcf24559362a6c0e14766a5cc2145bbd2" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.886909 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z7rjk"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.889611 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z7rjk"] Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.898345 4667 scope.go:117] "RemoveContainer" containerID="880a0bac667d6aed8e825340494bc387104a515d72ad0666735e0553de4c9a17" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.909967 4667 scope.go:117] "RemoveContainer" containerID="a626d823b2ec37753ecfaf1d3789049003e1bb7ad49a95702259a1c33a637049" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.919536 4667 scope.go:117] "RemoveContainer" containerID="d6efad009de94276a9d0733d1f15f31a8bf612324fb895731346587f8e2d3fad" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.928384 4667 scope.go:117] "RemoveContainer" containerID="b0b0e73b415f6e42889ee1e787f542e230cf95e5d1097dada75ffb332e916dfe" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.940008 4667 scope.go:117] "RemoveContainer" containerID="c34c22e95616f9c2ce252a9b85f28f986dd7dad21ad511d055e7f06add83fd83" Sep 29 17:13:11 crc kubenswrapper[4667]: I0929 17:13:11.953700 4667 scope.go:117] "RemoveContainer" containerID="7d7b6840ebde857f7144cdbb1b87e57d10422b41a1137a397ae73180f651c098" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.353975 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g5zzk"] Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354198 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56908f94-88ac-4298-98e5-3ba1315bd4ae" containerName="marketplace-operator" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354215 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="56908f94-88ac-4298-98e5-3ba1315bd4ae" containerName="marketplace-operator" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354223 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerName="extract-utilities" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354230 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerName="extract-utilities" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354239 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerName="extract-content" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354245 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerName="extract-content" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354253 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354259 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354268 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerName="extract-utilities" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354274 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerName="extract-utilities" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354282 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerName="extract-utilities" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354292 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerName="extract-utilities" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354309 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerName="extract-content" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354314 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerName="extract-content" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354328 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerName="extract-content" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354333 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerName="extract-content" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354340 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354346 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354352 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerName="extract-utilities" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354359 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerName="extract-utilities" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354365 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerName="extract-content" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354371 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerName="extract-content" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354380 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354386 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: E0929 17:13:12.354393 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354399 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354493 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354506 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="56908f94-88ac-4298-98e5-3ba1315bd4ae" containerName="marketplace-operator" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354513 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354519 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="3acdff5b-96b2-4100-a595-3c17295f9ef6" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.354526 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" containerName="registry-server" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.355313 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.357138 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.361987 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5zzk"] Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.407932 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a801d2-773e-4acd-905d-25dd9e9e2dcd-utilities\") pod \"certified-operators-g5zzk\" (UID: \"b0a801d2-773e-4acd-905d-25dd9e9e2dcd\") " pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.407974 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a801d2-773e-4acd-905d-25dd9e9e2dcd-catalog-content\") pod \"certified-operators-g5zzk\" (UID: \"b0a801d2-773e-4acd-905d-25dd9e9e2dcd\") " pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.408036 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t59pd\" (UniqueName: \"kubernetes.io/projected/b0a801d2-773e-4acd-905d-25dd9e9e2dcd-kube-api-access-t59pd\") pod \"certified-operators-g5zzk\" (UID: \"b0a801d2-773e-4acd-905d-25dd9e9e2dcd\") " pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.508702 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a801d2-773e-4acd-905d-25dd9e9e2dcd-utilities\") pod \"certified-operators-g5zzk\" (UID: \"b0a801d2-773e-4acd-905d-25dd9e9e2dcd\") " pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.508756 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a801d2-773e-4acd-905d-25dd9e9e2dcd-catalog-content\") pod \"certified-operators-g5zzk\" (UID: \"b0a801d2-773e-4acd-905d-25dd9e9e2dcd\") " pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.508796 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t59pd\" (UniqueName: \"kubernetes.io/projected/b0a801d2-773e-4acd-905d-25dd9e9e2dcd-kube-api-access-t59pd\") pod \"certified-operators-g5zzk\" (UID: \"b0a801d2-773e-4acd-905d-25dd9e9e2dcd\") " pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.509257 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0a801d2-773e-4acd-905d-25dd9e9e2dcd-catalog-content\") pod \"certified-operators-g5zzk\" (UID: \"b0a801d2-773e-4acd-905d-25dd9e9e2dcd\") " pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.509284 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0a801d2-773e-4acd-905d-25dd9e9e2dcd-utilities\") pod \"certified-operators-g5zzk\" (UID: \"b0a801d2-773e-4acd-905d-25dd9e9e2dcd\") " pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.524495 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t59pd\" (UniqueName: \"kubernetes.io/projected/b0a801d2-773e-4acd-905d-25dd9e9e2dcd-kube-api-access-t59pd\") pod \"certified-operators-g5zzk\" (UID: \"b0a801d2-773e-4acd-905d-25dd9e9e2dcd\") " pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.670786 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:12 crc kubenswrapper[4667]: I0929 17:13:12.998196 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5zzk"] Sep 29 17:13:13 crc kubenswrapper[4667]: W0929 17:13:13.004528 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0a801d2_773e_4acd_905d_25dd9e9e2dcd.slice/crio-b6cc6722806e4428addb64250813cdda40c2684b54be3c9bceda4093bafd093b WatchSource:0}: Error finding container b6cc6722806e4428addb64250813cdda40c2684b54be3c9bceda4093bafd093b: Status 404 returned error can't find the container with id b6cc6722806e4428addb64250813cdda40c2684b54be3c9bceda4093bafd093b Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.355634 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hhnh5"] Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.356435 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.358203 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.364684 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhnh5"] Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.418166 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcdb1214-877b-4b40-9e20-749d9d50c808-catalog-content\") pod \"redhat-marketplace-hhnh5\" (UID: \"fcdb1214-877b-4b40-9e20-749d9d50c808\") " pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.418218 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcdb1214-877b-4b40-9e20-749d9d50c808-utilities\") pod \"redhat-marketplace-hhnh5\" (UID: \"fcdb1214-877b-4b40-9e20-749d9d50c808\") " pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.418245 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkd6h\" (UniqueName: \"kubernetes.io/projected/fcdb1214-877b-4b40-9e20-749d9d50c808-kube-api-access-dkd6h\") pod \"redhat-marketplace-hhnh5\" (UID: \"fcdb1214-877b-4b40-9e20-749d9d50c808\") " pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.519209 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcdb1214-877b-4b40-9e20-749d9d50c808-utilities\") pod \"redhat-marketplace-hhnh5\" (UID: \"fcdb1214-877b-4b40-9e20-749d9d50c808\") " pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.519403 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkd6h\" (UniqueName: \"kubernetes.io/projected/fcdb1214-877b-4b40-9e20-749d9d50c808-kube-api-access-dkd6h\") pod \"redhat-marketplace-hhnh5\" (UID: \"fcdb1214-877b-4b40-9e20-749d9d50c808\") " pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.519472 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcdb1214-877b-4b40-9e20-749d9d50c808-catalog-content\") pod \"redhat-marketplace-hhnh5\" (UID: \"fcdb1214-877b-4b40-9e20-749d9d50c808\") " pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.519581 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcdb1214-877b-4b40-9e20-749d9d50c808-utilities\") pod \"redhat-marketplace-hhnh5\" (UID: \"fcdb1214-877b-4b40-9e20-749d9d50c808\") " pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.519787 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcdb1214-877b-4b40-9e20-749d9d50c808-catalog-content\") pod \"redhat-marketplace-hhnh5\" (UID: \"fcdb1214-877b-4b40-9e20-749d9d50c808\") " pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.534105 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkd6h\" (UniqueName: \"kubernetes.io/projected/fcdb1214-877b-4b40-9e20-749d9d50c808-kube-api-access-dkd6h\") pod \"redhat-marketplace-hhnh5\" (UID: \"fcdb1214-877b-4b40-9e20-749d9d50c808\") " pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.674890 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.805313 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhnh5"] Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.812973 4667 generic.go:334] "Generic (PLEG): container finished" podID="b0a801d2-773e-4acd-905d-25dd9e9e2dcd" containerID="6d94898b622e357edcfe8251db0f615cd96ba989b8fabfbb5818b2d07df58982" exitCode=0 Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.813049 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5zzk" event={"ID":"b0a801d2-773e-4acd-905d-25dd9e9e2dcd","Type":"ContainerDied","Data":"6d94898b622e357edcfe8251db0f615cd96ba989b8fabfbb5818b2d07df58982"} Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.813091 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5zzk" event={"ID":"b0a801d2-773e-4acd-905d-25dd9e9e2dcd","Type":"ContainerStarted","Data":"b6cc6722806e4428addb64250813cdda40c2684b54be3c9bceda4093bafd093b"} Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.820534 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3795f5d0-58d5-40a1-bfcb-98da8a26e905" path="/var/lib/kubelet/pods/3795f5d0-58d5-40a1-bfcb-98da8a26e905/volumes" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.821099 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3acdff5b-96b2-4100-a595-3c17295f9ef6" path="/var/lib/kubelet/pods/3acdff5b-96b2-4100-a595-3c17295f9ef6/volumes" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.821689 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56908f94-88ac-4298-98e5-3ba1315bd4ae" path="/var/lib/kubelet/pods/56908f94-88ac-4298-98e5-3ba1315bd4ae/volumes" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.822960 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af6b10ff-6922-44fc-a489-514cbcffe69d" path="/var/lib/kubelet/pods/af6b10ff-6922-44fc-a489-514cbcffe69d/volumes" Sep 29 17:13:13 crc kubenswrapper[4667]: I0929 17:13:13.834180 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b" path="/var/lib/kubelet/pods/eb7bdd67-2d8f-4f06-96bb-64fd7f6be71b/volumes" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.752342 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7jzk2"] Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.753594 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.755493 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.764330 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jzk2"] Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.818390 4667 generic.go:334] "Generic (PLEG): container finished" podID="b0a801d2-773e-4acd-905d-25dd9e9e2dcd" containerID="521b6e1275ad692af7f80ebba6e31bfcfdbb1afcebe6b84ae8e58fab4a70f509" exitCode=0 Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.818466 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5zzk" event={"ID":"b0a801d2-773e-4acd-905d-25dd9e9e2dcd","Type":"ContainerDied","Data":"521b6e1275ad692af7f80ebba6e31bfcfdbb1afcebe6b84ae8e58fab4a70f509"} Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.819914 4667 generic.go:334] "Generic (PLEG): container finished" podID="fcdb1214-877b-4b40-9e20-749d9d50c808" containerID="3b48cefdb28cf5b51de20b9a677dd6a38aad66e00b88cd659a8a0f8c85455876" exitCode=0 Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.819947 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhnh5" event={"ID":"fcdb1214-877b-4b40-9e20-749d9d50c808","Type":"ContainerDied","Data":"3b48cefdb28cf5b51de20b9a677dd6a38aad66e00b88cd659a8a0f8c85455876"} Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.819968 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhnh5" event={"ID":"fcdb1214-877b-4b40-9e20-749d9d50c808","Type":"ContainerStarted","Data":"0b5afc60a1e97438c4a635f88d9ad5e3746c3295b654d2b88d93d9210e1888b7"} Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.836217 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg7lj\" (UniqueName: \"kubernetes.io/projected/5877d4d8-3fbf-448c-963f-c6e493fb1c64-kube-api-access-kg7lj\") pod \"redhat-operators-7jzk2\" (UID: \"5877d4d8-3fbf-448c-963f-c6e493fb1c64\") " pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.836262 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5877d4d8-3fbf-448c-963f-c6e493fb1c64-catalog-content\") pod \"redhat-operators-7jzk2\" (UID: \"5877d4d8-3fbf-448c-963f-c6e493fb1c64\") " pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.836290 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5877d4d8-3fbf-448c-963f-c6e493fb1c64-utilities\") pod \"redhat-operators-7jzk2\" (UID: \"5877d4d8-3fbf-448c-963f-c6e493fb1c64\") " pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.937357 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg7lj\" (UniqueName: \"kubernetes.io/projected/5877d4d8-3fbf-448c-963f-c6e493fb1c64-kube-api-access-kg7lj\") pod \"redhat-operators-7jzk2\" (UID: \"5877d4d8-3fbf-448c-963f-c6e493fb1c64\") " pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.937392 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5877d4d8-3fbf-448c-963f-c6e493fb1c64-catalog-content\") pod \"redhat-operators-7jzk2\" (UID: \"5877d4d8-3fbf-448c-963f-c6e493fb1c64\") " pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.937413 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5877d4d8-3fbf-448c-963f-c6e493fb1c64-utilities\") pod \"redhat-operators-7jzk2\" (UID: \"5877d4d8-3fbf-448c-963f-c6e493fb1c64\") " pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.938034 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5877d4d8-3fbf-448c-963f-c6e493fb1c64-catalog-content\") pod \"redhat-operators-7jzk2\" (UID: \"5877d4d8-3fbf-448c-963f-c6e493fb1c64\") " pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.938079 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5877d4d8-3fbf-448c-963f-c6e493fb1c64-utilities\") pod \"redhat-operators-7jzk2\" (UID: \"5877d4d8-3fbf-448c-963f-c6e493fb1c64\") " pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:14 crc kubenswrapper[4667]: I0929 17:13:14.952591 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg7lj\" (UniqueName: \"kubernetes.io/projected/5877d4d8-3fbf-448c-963f-c6e493fb1c64-kube-api-access-kg7lj\") pod \"redhat-operators-7jzk2\" (UID: \"5877d4d8-3fbf-448c-963f-c6e493fb1c64\") " pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.110570 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.452602 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jzk2"] Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.752958 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tcltq"] Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.754168 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.756824 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.762313 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tcltq"] Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.824976 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g5zzk" event={"ID":"b0a801d2-773e-4acd-905d-25dd9e9e2dcd","Type":"ContainerStarted","Data":"cb058b8b569548925eb42b50a84d7e87c02045bc5ed3b9525376d37cd881efa9"} Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.826603 4667 generic.go:334] "Generic (PLEG): container finished" podID="5877d4d8-3fbf-448c-963f-c6e493fb1c64" containerID="b8bda85cb56d490da4cafb41ee6c32ceeeda4797d41639e8d45ba2779bdf81e3" exitCode=0 Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.826639 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jzk2" event={"ID":"5877d4d8-3fbf-448c-963f-c6e493fb1c64","Type":"ContainerDied","Data":"b8bda85cb56d490da4cafb41ee6c32ceeeda4797d41639e8d45ba2779bdf81e3"} Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.826671 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jzk2" event={"ID":"5877d4d8-3fbf-448c-963f-c6e493fb1c64","Type":"ContainerStarted","Data":"9f72b6947f7d61cb416e62a9345d6c3d9be9ae325a6d20ec88f72bd8335ab551"} Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.828913 4667 generic.go:334] "Generic (PLEG): container finished" podID="fcdb1214-877b-4b40-9e20-749d9d50c808" containerID="b59406947ee8e81cb94f44bbb823a0a56d4c4253289f2529ebd0feb034429d6a" exitCode=0 Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.828954 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhnh5" event={"ID":"fcdb1214-877b-4b40-9e20-749d9d50c808","Type":"ContainerDied","Data":"b59406947ee8e81cb94f44bbb823a0a56d4c4253289f2529ebd0feb034429d6a"} Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.840972 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g5zzk" podStartSLOduration=2.113832788 podStartE2EDuration="3.84095769s" podCreationTimestamp="2025-09-29 17:13:12 +0000 UTC" firstStartedPulling="2025-09-29 17:13:13.814932746 +0000 UTC m=+222.312779515" lastFinishedPulling="2025-09-29 17:13:15.542057647 +0000 UTC m=+224.039904417" observedRunningTime="2025-09-29 17:13:15.839709256 +0000 UTC m=+224.337556025" watchObservedRunningTime="2025-09-29 17:13:15.84095769 +0000 UTC m=+224.338804458" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.846731 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50fa2887-a6dd-4b50-bd32-cc65c5a380ce-catalog-content\") pod \"community-operators-tcltq\" (UID: \"50fa2887-a6dd-4b50-bd32-cc65c5a380ce\") " pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.846857 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lccdv\" (UniqueName: \"kubernetes.io/projected/50fa2887-a6dd-4b50-bd32-cc65c5a380ce-kube-api-access-lccdv\") pod \"community-operators-tcltq\" (UID: \"50fa2887-a6dd-4b50-bd32-cc65c5a380ce\") " pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.846883 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50fa2887-a6dd-4b50-bd32-cc65c5a380ce-utilities\") pod \"community-operators-tcltq\" (UID: \"50fa2887-a6dd-4b50-bd32-cc65c5a380ce\") " pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.947323 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50fa2887-a6dd-4b50-bd32-cc65c5a380ce-catalog-content\") pod \"community-operators-tcltq\" (UID: \"50fa2887-a6dd-4b50-bd32-cc65c5a380ce\") " pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.947387 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lccdv\" (UniqueName: \"kubernetes.io/projected/50fa2887-a6dd-4b50-bd32-cc65c5a380ce-kube-api-access-lccdv\") pod \"community-operators-tcltq\" (UID: \"50fa2887-a6dd-4b50-bd32-cc65c5a380ce\") " pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.947409 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50fa2887-a6dd-4b50-bd32-cc65c5a380ce-utilities\") pod \"community-operators-tcltq\" (UID: \"50fa2887-a6dd-4b50-bd32-cc65c5a380ce\") " pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.947801 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50fa2887-a6dd-4b50-bd32-cc65c5a380ce-catalog-content\") pod \"community-operators-tcltq\" (UID: \"50fa2887-a6dd-4b50-bd32-cc65c5a380ce\") " pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.947819 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50fa2887-a6dd-4b50-bd32-cc65c5a380ce-utilities\") pod \"community-operators-tcltq\" (UID: \"50fa2887-a6dd-4b50-bd32-cc65c5a380ce\") " pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:15 crc kubenswrapper[4667]: I0929 17:13:15.962117 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lccdv\" (UniqueName: \"kubernetes.io/projected/50fa2887-a6dd-4b50-bd32-cc65c5a380ce-kube-api-access-lccdv\") pod \"community-operators-tcltq\" (UID: \"50fa2887-a6dd-4b50-bd32-cc65c5a380ce\") " pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:16 crc kubenswrapper[4667]: I0929 17:13:16.079498 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:16 crc kubenswrapper[4667]: I0929 17:13:16.428464 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tcltq"] Sep 29 17:13:16 crc kubenswrapper[4667]: W0929 17:13:16.431771 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50fa2887_a6dd_4b50_bd32_cc65c5a380ce.slice/crio-634ace9219bfa854282ce26a013a2b4af124f5d4b69b5b1d3fcc1d1d45b7904c WatchSource:0}: Error finding container 634ace9219bfa854282ce26a013a2b4af124f5d4b69b5b1d3fcc1d1d45b7904c: Status 404 returned error can't find the container with id 634ace9219bfa854282ce26a013a2b4af124f5d4b69b5b1d3fcc1d1d45b7904c Sep 29 17:13:16 crc kubenswrapper[4667]: I0929 17:13:16.834274 4667 generic.go:334] "Generic (PLEG): container finished" podID="50fa2887-a6dd-4b50-bd32-cc65c5a380ce" containerID="83d2158118f194e9dc99aa546b5ee69f5e1c270535487402b9b511950ac057eb" exitCode=0 Sep 29 17:13:16 crc kubenswrapper[4667]: I0929 17:13:16.834370 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcltq" event={"ID":"50fa2887-a6dd-4b50-bd32-cc65c5a380ce","Type":"ContainerDied","Data":"83d2158118f194e9dc99aa546b5ee69f5e1c270535487402b9b511950ac057eb"} Sep 29 17:13:16 crc kubenswrapper[4667]: I0929 17:13:16.834717 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcltq" event={"ID":"50fa2887-a6dd-4b50-bd32-cc65c5a380ce","Type":"ContainerStarted","Data":"634ace9219bfa854282ce26a013a2b4af124f5d4b69b5b1d3fcc1d1d45b7904c"} Sep 29 17:13:16 crc kubenswrapper[4667]: I0929 17:13:16.838450 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhnh5" event={"ID":"fcdb1214-877b-4b40-9e20-749d9d50c808","Type":"ContainerStarted","Data":"7b46d03610d8a863ef7c0f7ce0eb1cf6997be6a6e7bf33d76ac99188ffbf50e2"} Sep 29 17:13:16 crc kubenswrapper[4667]: I0929 17:13:16.847390 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jzk2" event={"ID":"5877d4d8-3fbf-448c-963f-c6e493fb1c64","Type":"ContainerStarted","Data":"6290919a63dd922a97623e13cab6930402f8ff3790ef1988fcf03d024f865050"} Sep 29 17:13:16 crc kubenswrapper[4667]: I0929 17:13:16.880491 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hhnh5" podStartSLOduration=2.094673318 podStartE2EDuration="3.880461763s" podCreationTimestamp="2025-09-29 17:13:13 +0000 UTC" firstStartedPulling="2025-09-29 17:13:14.820976301 +0000 UTC m=+223.318823070" lastFinishedPulling="2025-09-29 17:13:16.606764746 +0000 UTC m=+225.104611515" observedRunningTime="2025-09-29 17:13:16.880254869 +0000 UTC m=+225.378101638" watchObservedRunningTime="2025-09-29 17:13:16.880461763 +0000 UTC m=+225.378308532" Sep 29 17:13:17 crc kubenswrapper[4667]: I0929 17:13:17.854248 4667 generic.go:334] "Generic (PLEG): container finished" podID="50fa2887-a6dd-4b50-bd32-cc65c5a380ce" containerID="06e333159578d12a7fffed82339df81fedf5d9fa32ba422a737a4a3273efb4d1" exitCode=0 Sep 29 17:13:17 crc kubenswrapper[4667]: I0929 17:13:17.854291 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcltq" event={"ID":"50fa2887-a6dd-4b50-bd32-cc65c5a380ce","Type":"ContainerDied","Data":"06e333159578d12a7fffed82339df81fedf5d9fa32ba422a737a4a3273efb4d1"} Sep 29 17:13:17 crc kubenswrapper[4667]: I0929 17:13:17.858254 4667 generic.go:334] "Generic (PLEG): container finished" podID="5877d4d8-3fbf-448c-963f-c6e493fb1c64" containerID="6290919a63dd922a97623e13cab6930402f8ff3790ef1988fcf03d024f865050" exitCode=0 Sep 29 17:13:17 crc kubenswrapper[4667]: I0929 17:13:17.858292 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jzk2" event={"ID":"5877d4d8-3fbf-448c-963f-c6e493fb1c64","Type":"ContainerDied","Data":"6290919a63dd922a97623e13cab6930402f8ff3790ef1988fcf03d024f865050"} Sep 29 17:13:18 crc kubenswrapper[4667]: I0929 17:13:18.863555 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jzk2" event={"ID":"5877d4d8-3fbf-448c-963f-c6e493fb1c64","Type":"ContainerStarted","Data":"60e0d54bea6fe290c271cc9dd50e1a3f6152d5dba3be391c84c2200c0e02ea9f"} Sep 29 17:13:18 crc kubenswrapper[4667]: I0929 17:13:18.865359 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcltq" event={"ID":"50fa2887-a6dd-4b50-bd32-cc65c5a380ce","Type":"ContainerStarted","Data":"f46781031f4ded68551c65755d4a749f95bbac2e5261c9b53c97c00a3e95162b"} Sep 29 17:13:18 crc kubenswrapper[4667]: I0929 17:13:18.879994 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7jzk2" podStartSLOduration=2.291974685 podStartE2EDuration="4.879981745s" podCreationTimestamp="2025-09-29 17:13:14 +0000 UTC" firstStartedPulling="2025-09-29 17:13:15.827579881 +0000 UTC m=+224.325426650" lastFinishedPulling="2025-09-29 17:13:18.41558694 +0000 UTC m=+226.913433710" observedRunningTime="2025-09-29 17:13:18.877625966 +0000 UTC m=+227.375472735" watchObservedRunningTime="2025-09-29 17:13:18.879981745 +0000 UTC m=+227.377828514" Sep 29 17:13:18 crc kubenswrapper[4667]: I0929 17:13:18.891756 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tcltq" podStartSLOduration=2.210831958 podStartE2EDuration="3.891744102s" podCreationTimestamp="2025-09-29 17:13:15 +0000 UTC" firstStartedPulling="2025-09-29 17:13:16.835233227 +0000 UTC m=+225.333079996" lastFinishedPulling="2025-09-29 17:13:18.516145371 +0000 UTC m=+227.013992140" observedRunningTime="2025-09-29 17:13:18.889457986 +0000 UTC m=+227.387304754" watchObservedRunningTime="2025-09-29 17:13:18.891744102 +0000 UTC m=+227.389590871" Sep 29 17:13:22 crc kubenswrapper[4667]: I0929 17:13:22.672041 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:22 crc kubenswrapper[4667]: I0929 17:13:22.672373 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:22 crc kubenswrapper[4667]: I0929 17:13:22.698552 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:22 crc kubenswrapper[4667]: I0929 17:13:22.904316 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g5zzk" Sep 29 17:13:23 crc kubenswrapper[4667]: I0929 17:13:23.675274 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:23 crc kubenswrapper[4667]: I0929 17:13:23.675467 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:23 crc kubenswrapper[4667]: I0929 17:13:23.702099 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:23 crc kubenswrapper[4667]: I0929 17:13:23.908086 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hhnh5" Sep 29 17:13:25 crc kubenswrapper[4667]: I0929 17:13:25.111129 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:25 crc kubenswrapper[4667]: I0929 17:13:25.111271 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:25 crc kubenswrapper[4667]: I0929 17:13:25.137469 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:25 crc kubenswrapper[4667]: I0929 17:13:25.918093 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7jzk2" Sep 29 17:13:26 crc kubenswrapper[4667]: I0929 17:13:26.080160 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:26 crc kubenswrapper[4667]: I0929 17:13:26.080195 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:26 crc kubenswrapper[4667]: I0929 17:13:26.104682 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:26 crc kubenswrapper[4667]: I0929 17:13:26.922239 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tcltq" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.799357 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw"] Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.800220 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.801912 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.801942 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.802133 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.802359 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.802998 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.806958 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw"] Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.876031 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f739a92-fd0f-4675-adfc-f4de5b44a914-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-bgjmw\" (UID: \"6f739a92-fd0f-4675-adfc-f4de5b44a914\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.876278 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/6f739a92-fd0f-4675-adfc-f4de5b44a914-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-bgjmw\" (UID: \"6f739a92-fd0f-4675-adfc-f4de5b44a914\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.876300 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzt9r\" (UniqueName: \"kubernetes.io/projected/6f739a92-fd0f-4675-adfc-f4de5b44a914-kube-api-access-vzt9r\") pod \"cluster-monitoring-operator-6d5b84845-bgjmw\" (UID: \"6f739a92-fd0f-4675-adfc-f4de5b44a914\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.977150 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/6f739a92-fd0f-4675-adfc-f4de5b44a914-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-bgjmw\" (UID: \"6f739a92-fd0f-4675-adfc-f4de5b44a914\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.977181 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzt9r\" (UniqueName: \"kubernetes.io/projected/6f739a92-fd0f-4675-adfc-f4de5b44a914-kube-api-access-vzt9r\") pod \"cluster-monitoring-operator-6d5b84845-bgjmw\" (UID: \"6f739a92-fd0f-4675-adfc-f4de5b44a914\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.977237 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f739a92-fd0f-4675-adfc-f4de5b44a914-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-bgjmw\" (UID: \"6f739a92-fd0f-4675-adfc-f4de5b44a914\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.978024 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/6f739a92-fd0f-4675-adfc-f4de5b44a914-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-bgjmw\" (UID: \"6f739a92-fd0f-4675-adfc-f4de5b44a914\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.981592 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/6f739a92-fd0f-4675-adfc-f4de5b44a914-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-bgjmw\" (UID: \"6f739a92-fd0f-4675-adfc-f4de5b44a914\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:39 crc kubenswrapper[4667]: I0929 17:13:39.989763 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzt9r\" (UniqueName: \"kubernetes.io/projected/6f739a92-fd0f-4675-adfc-f4de5b44a914-kube-api-access-vzt9r\") pod \"cluster-monitoring-operator-6d5b84845-bgjmw\" (UID: \"6f739a92-fd0f-4675-adfc-f4de5b44a914\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:40 crc kubenswrapper[4667]: I0929 17:13:40.112074 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" Sep 29 17:13:40 crc kubenswrapper[4667]: I0929 17:13:40.440250 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw"] Sep 29 17:13:40 crc kubenswrapper[4667]: W0929 17:13:40.444185 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f739a92_fd0f_4675_adfc_f4de5b44a914.slice/crio-aaeced5ad2da2c2b493276b04dba65bcd748adce90ffb6417a9d0442958cbab7 WatchSource:0}: Error finding container aaeced5ad2da2c2b493276b04dba65bcd748adce90ffb6417a9d0442958cbab7: Status 404 returned error can't find the container with id aaeced5ad2da2c2b493276b04dba65bcd748adce90ffb6417a9d0442958cbab7 Sep 29 17:13:40 crc kubenswrapper[4667]: I0929 17:13:40.944126 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" event={"ID":"6f739a92-fd0f-4675-adfc-f4de5b44a914","Type":"ContainerStarted","Data":"aaeced5ad2da2c2b493276b04dba65bcd748adce90ffb6417a9d0442958cbab7"} Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.300961 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xmbzg"] Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.302277 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.309494 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xmbzg"] Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.329495 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b37a4fc8-6c64-402b-8b27-d3ce92271b26-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.329548 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b37a4fc8-6c64-402b-8b27-d3ce92271b26-registry-tls\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.329573 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b37a4fc8-6c64-402b-8b27-d3ce92271b26-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.329593 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b37a4fc8-6c64-402b-8b27-d3ce92271b26-bound-sa-token\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.329651 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.329946 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv4zd\" (UniqueName: \"kubernetes.io/projected/b37a4fc8-6c64-402b-8b27-d3ce92271b26-kube-api-access-dv4zd\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.329975 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b37a4fc8-6c64-402b-8b27-d3ce92271b26-trusted-ca\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.330026 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b37a4fc8-6c64-402b-8b27-d3ce92271b26-registry-certificates\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.362145 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.393914 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b"] Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.394497 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.399446 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-h9grl" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.399451 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.401942 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b"] Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.430872 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv4zd\" (UniqueName: \"kubernetes.io/projected/b37a4fc8-6c64-402b-8b27-d3ce92271b26-kube-api-access-dv4zd\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.430903 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b37a4fc8-6c64-402b-8b27-d3ce92271b26-trusted-ca\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.430930 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b37a4fc8-6c64-402b-8b27-d3ce92271b26-registry-certificates\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.430972 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b37a4fc8-6c64-402b-8b27-d3ce92271b26-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.430995 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b37a4fc8-6c64-402b-8b27-d3ce92271b26-registry-tls\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.431013 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b37a4fc8-6c64-402b-8b27-d3ce92271b26-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.431030 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b37a4fc8-6c64-402b-8b27-d3ce92271b26-bound-sa-token\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.431066 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/a3b8df09-3f2b-4a0f-ae1b-879092cba994-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-8fn2b\" (UID: \"a3b8df09-3f2b-4a0f-ae1b-879092cba994\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.431445 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b37a4fc8-6c64-402b-8b27-d3ce92271b26-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.432035 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b37a4fc8-6c64-402b-8b27-d3ce92271b26-trusted-ca\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.432130 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b37a4fc8-6c64-402b-8b27-d3ce92271b26-registry-certificates\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.436452 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b37a4fc8-6c64-402b-8b27-d3ce92271b26-registry-tls\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.439246 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b37a4fc8-6c64-402b-8b27-d3ce92271b26-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.446363 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv4zd\" (UniqueName: \"kubernetes.io/projected/b37a4fc8-6c64-402b-8b27-d3ce92271b26-kube-api-access-dv4zd\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.459517 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b37a4fc8-6c64-402b-8b27-d3ce92271b26-bound-sa-token\") pod \"image-registry-66df7c8f76-xmbzg\" (UID: \"b37a4fc8-6c64-402b-8b27-d3ce92271b26\") " pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.531856 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/a3b8df09-3f2b-4a0f-ae1b-879092cba994-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-8fn2b\" (UID: \"a3b8df09-3f2b-4a0f-ae1b-879092cba994\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" Sep 29 17:13:44 crc kubenswrapper[4667]: E0929 17:13:44.532000 4667 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Sep 29 17:13:44 crc kubenswrapper[4667]: E0929 17:13:44.532074 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3b8df09-3f2b-4a0f-ae1b-879092cba994-tls-certificates podName:a3b8df09-3f2b-4a0f-ae1b-879092cba994 nodeName:}" failed. No retries permitted until 2025-09-29 17:13:45.032048187 +0000 UTC m=+253.529894957 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/a3b8df09-3f2b-4a0f-ae1b-879092cba994-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-8fn2b" (UID: "a3b8df09-3f2b-4a0f-ae1b-879092cba994") : secret "prometheus-operator-admission-webhook-tls" not found Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.620348 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.959415 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xmbzg"] Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.968001 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" event={"ID":"6f739a92-fd0f-4675-adfc-f4de5b44a914","Type":"ContainerStarted","Data":"b27bad3e3f75e98176a3be5d628f38fdd08554897f45f8e32eaa819b673d2cc1"} Sep 29 17:13:44 crc kubenswrapper[4667]: W0929 17:13:44.970434 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb37a4fc8_6c64_402b_8b27_d3ce92271b26.slice/crio-082c7512ad05ee43a726f264c28c4260fe304ff43840bb9fc5b3acf1e8c7d72b WatchSource:0}: Error finding container 082c7512ad05ee43a726f264c28c4260fe304ff43840bb9fc5b3acf1e8c7d72b: Status 404 returned error can't find the container with id 082c7512ad05ee43a726f264c28c4260fe304ff43840bb9fc5b3acf1e8c7d72b Sep 29 17:13:44 crc kubenswrapper[4667]: I0929 17:13:44.996516 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-bgjmw" podStartSLOduration=2.495576564 podStartE2EDuration="5.996503156s" podCreationTimestamp="2025-09-29 17:13:39 +0000 UTC" firstStartedPulling="2025-09-29 17:13:40.445904589 +0000 UTC m=+248.943751357" lastFinishedPulling="2025-09-29 17:13:43.946831179 +0000 UTC m=+252.444677949" observedRunningTime="2025-09-29 17:13:44.995614738 +0000 UTC m=+253.493461507" watchObservedRunningTime="2025-09-29 17:13:44.996503156 +0000 UTC m=+253.494349926" Sep 29 17:13:45 crc kubenswrapper[4667]: I0929 17:13:45.037344 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/a3b8df09-3f2b-4a0f-ae1b-879092cba994-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-8fn2b\" (UID: \"a3b8df09-3f2b-4a0f-ae1b-879092cba994\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" Sep 29 17:13:45 crc kubenswrapper[4667]: I0929 17:13:45.040585 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/a3b8df09-3f2b-4a0f-ae1b-879092cba994-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-8fn2b\" (UID: \"a3b8df09-3f2b-4a0f-ae1b-879092cba994\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" Sep 29 17:13:45 crc kubenswrapper[4667]: I0929 17:13:45.305868 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" Sep 29 17:13:45 crc kubenswrapper[4667]: I0929 17:13:45.433293 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b"] Sep 29 17:13:45 crc kubenswrapper[4667]: W0929 17:13:45.438768 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3b8df09_3f2b_4a0f_ae1b_879092cba994.slice/crio-9a2dcdddaf70a3a940991d31bcd79bc77cfda98cd3d40f63a5e5d433fb3d8413 WatchSource:0}: Error finding container 9a2dcdddaf70a3a940991d31bcd79bc77cfda98cd3d40f63a5e5d433fb3d8413: Status 404 returned error can't find the container with id 9a2dcdddaf70a3a940991d31bcd79bc77cfda98cd3d40f63a5e5d433fb3d8413 Sep 29 17:13:45 crc kubenswrapper[4667]: I0929 17:13:45.972374 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" event={"ID":"a3b8df09-3f2b-4a0f-ae1b-879092cba994","Type":"ContainerStarted","Data":"9a2dcdddaf70a3a940991d31bcd79bc77cfda98cd3d40f63a5e5d433fb3d8413"} Sep 29 17:13:45 crc kubenswrapper[4667]: I0929 17:13:45.973901 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" event={"ID":"b37a4fc8-6c64-402b-8b27-d3ce92271b26","Type":"ContainerStarted","Data":"2ae0755502c6737c86fdbf4b50d84782c8f2dce38246ef208cc3463162840617"} Sep 29 17:13:45 crc kubenswrapper[4667]: I0929 17:13:45.973925 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:13:45 crc kubenswrapper[4667]: I0929 17:13:45.973935 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" event={"ID":"b37a4fc8-6c64-402b-8b27-d3ce92271b26","Type":"ContainerStarted","Data":"082c7512ad05ee43a726f264c28c4260fe304ff43840bb9fc5b3acf1e8c7d72b"} Sep 29 17:13:45 crc kubenswrapper[4667]: I0929 17:13:45.986935 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" podStartSLOduration=1.9869255780000001 podStartE2EDuration="1.986925578s" podCreationTimestamp="2025-09-29 17:13:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:13:45.985041395 +0000 UTC m=+254.482888163" watchObservedRunningTime="2025-09-29 17:13:45.986925578 +0000 UTC m=+254.484772336" Sep 29 17:13:47 crc kubenswrapper[4667]: I0929 17:13:47.982490 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" event={"ID":"a3b8df09-3f2b-4a0f-ae1b-879092cba994","Type":"ContainerStarted","Data":"129b070be4982d7f02b5f6107ad2e7ae923e11fee7f528056cb0fad38166a918"} Sep 29 17:13:47 crc kubenswrapper[4667]: I0929 17:13:47.982676 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" Sep 29 17:13:47 crc kubenswrapper[4667]: I0929 17:13:47.985936 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" Sep 29 17:13:47 crc kubenswrapper[4667]: I0929 17:13:47.993967 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-8fn2b" podStartSLOduration=1.785281825 podStartE2EDuration="3.993956512s" podCreationTimestamp="2025-09-29 17:13:44 +0000 UTC" firstStartedPulling="2025-09-29 17:13:45.44049278 +0000 UTC m=+253.938339549" lastFinishedPulling="2025-09-29 17:13:47.649167468 +0000 UTC m=+256.147014236" observedRunningTime="2025-09-29 17:13:47.992230681 +0000 UTC m=+256.490077450" watchObservedRunningTime="2025-09-29 17:13:47.993956512 +0000 UTC m=+256.491803281" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.427448 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-vg2kc"] Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.428139 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.429336 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.429602 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-qvhzs" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.429663 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.429806 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.436045 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-vg2kc"] Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.574453 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.574504 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrdtn\" (UniqueName: \"kubernetes.io/projected/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-kube-api-access-rrdtn\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.574551 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.574689 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-metrics-client-ca\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.675601 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.675655 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrdtn\" (UniqueName: \"kubernetes.io/projected/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-kube-api-access-rrdtn\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.675702 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.675730 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-metrics-client-ca\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.676525 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-metrics-client-ca\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.679992 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.680515 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.688511 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrdtn\" (UniqueName: \"kubernetes.io/projected/4fec1db8-a0ef-4556-8bf4-02f5d5266fc6-kube-api-access-rrdtn\") pod \"prometheus-operator-db54df47d-vg2kc\" (UID: \"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6\") " pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:48 crc kubenswrapper[4667]: I0929 17:13:48.738693 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" Sep 29 17:13:49 crc kubenswrapper[4667]: I0929 17:13:49.068143 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-vg2kc"] Sep 29 17:13:49 crc kubenswrapper[4667]: W0929 17:13:49.072816 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fec1db8_a0ef_4556_8bf4_02f5d5266fc6.slice/crio-2d245d101c39b567624cf6eb6a996b8138b7467c032d77fff86fb7c9b87efce9 WatchSource:0}: Error finding container 2d245d101c39b567624cf6eb6a996b8138b7467c032d77fff86fb7c9b87efce9: Status 404 returned error can't find the container with id 2d245d101c39b567624cf6eb6a996b8138b7467c032d77fff86fb7c9b87efce9 Sep 29 17:13:49 crc kubenswrapper[4667]: I0929 17:13:49.990479 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" event={"ID":"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6","Type":"ContainerStarted","Data":"2d245d101c39b567624cf6eb6a996b8138b7467c032d77fff86fb7c9b87efce9"} Sep 29 17:13:51 crc kubenswrapper[4667]: I0929 17:13:51.999238 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" event={"ID":"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6","Type":"ContainerStarted","Data":"cb0ee40deab648a46adb098de90d82465522c206c035f2486f7c727d07b5c01e"} Sep 29 17:13:53 crc kubenswrapper[4667]: I0929 17:13:53.003912 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" event={"ID":"4fec1db8-a0ef-4556-8bf4-02f5d5266fc6","Type":"ContainerStarted","Data":"e7d7aa8d30c6a329f761d18827985d322edfad9847e906c53d31d877e0fa59bf"} Sep 29 17:13:53 crc kubenswrapper[4667]: I0929 17:13:53.015360 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-vg2kc" podStartSLOduration=2.206963047 podStartE2EDuration="5.015346941s" podCreationTimestamp="2025-09-29 17:13:48 +0000 UTC" firstStartedPulling="2025-09-29 17:13:49.074423279 +0000 UTC m=+257.572270049" lastFinishedPulling="2025-09-29 17:13:51.882807174 +0000 UTC m=+260.380653943" observedRunningTime="2025-09-29 17:13:53.015106553 +0000 UTC m=+261.512953323" watchObservedRunningTime="2025-09-29 17:13:53.015346941 +0000 UTC m=+261.513193709" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.723011 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj"] Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.724024 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.727438 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.727941 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-4nl6f" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.728026 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.736015 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-nxqpg"] Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.736908 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.738462 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj"] Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.739504 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.739642 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.741338 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-772jb" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.747704 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv"] Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.748381 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: W0929 17:13:54.749247 4667 reflector.go:561] object-"openshift-monitoring"/"kube-state-metrics-tls": failed to list *v1.Secret: secrets "kube-state-metrics-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-monitoring": no relationship found between node 'crc' and this object Sep 29 17:13:54 crc kubenswrapper[4667]: E0929 17:13:54.749288 4667 reflector.go:158] "Unhandled Error" err="object-\"openshift-monitoring\"/\"kube-state-metrics-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"kube-state-metrics-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-monitoring\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 29 17:13:54 crc kubenswrapper[4667]: W0929 17:13:54.749374 4667 reflector.go:561] object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config": failed to list *v1.Secret: secrets "kube-state-metrics-kube-rbac-proxy-config" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-monitoring": no relationship found between node 'crc' and this object Sep 29 17:13:54 crc kubenswrapper[4667]: E0929 17:13:54.749388 4667 reflector.go:158] "Unhandled Error" err="object-\"openshift-monitoring\"/\"kube-state-metrics-kube-rbac-proxy-config\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"kube-state-metrics-kube-rbac-proxy-config\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-monitoring\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 29 17:13:54 crc kubenswrapper[4667]: W0929 17:13:54.749466 4667 reflector.go:561] object-"openshift-monitoring"/"kube-state-metrics-dockercfg-9vm6b": failed to list *v1.Secret: secrets "kube-state-metrics-dockercfg-9vm6b" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-monitoring": no relationship found between node 'crc' and this object Sep 29 17:13:54 crc kubenswrapper[4667]: E0929 17:13:54.749484 4667 reflector.go:158] "Unhandled Error" err="object-\"openshift-monitoring\"/\"kube-state-metrics-dockercfg-9vm6b\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"kube-state-metrics-dockercfg-9vm6b\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-monitoring\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 29 17:13:54 crc kubenswrapper[4667]: W0929 17:13:54.750372 4667 reflector.go:561] object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap": failed to list *v1.ConfigMap: configmaps "kube-state-metrics-custom-resource-state-configmap" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-monitoring": no relationship found between node 'crc' and this object Sep 29 17:13:54 crc kubenswrapper[4667]: E0929 17:13:54.750396 4667 reflector.go:158] "Unhandled Error" err="object-\"openshift-monitoring\"/\"kube-state-metrics-custom-resource-state-configmap\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-state-metrics-custom-resource-state-configmap\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-monitoring\": no relationship found between node 'crc' and this object" logger="UnhandledError" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.752626 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-textfile\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.752657 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-tls\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.752679 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8f14efb1-7152-4798-af6a-7a60d8cacccf-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.752735 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/45fe7b22-d9bd-490b-9a11-3a967263ee66-root\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.752923 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/8f14efb1-7152-4798-af6a-7a60d8cacccf-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.752956 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/8f14efb1-7152-4798-af6a-7a60d8cacccf-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.752982 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/45fe7b22-d9bd-490b-9a11-3a967263ee66-sys\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.753003 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-wtmp\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.753056 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbpkj\" (UniqueName: \"kubernetes.io/projected/45fe7b22-d9bd-490b-9a11-3a967263ee66-kube-api-access-kbpkj\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.753085 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv567\" (UniqueName: \"kubernetes.io/projected/8f14efb1-7152-4798-af6a-7a60d8cacccf-kube-api-access-cv567\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.753117 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.753155 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/45fe7b22-d9bd-490b-9a11-3a967263ee66-metrics-client-ca\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.766774 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv"] Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.854395 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbpkj\" (UniqueName: \"kubernetes.io/projected/45fe7b22-d9bd-490b-9a11-3a967263ee66-kube-api-access-kbpkj\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.854780 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwg6j\" (UniqueName: \"kubernetes.io/projected/f0420681-49bc-466a-976e-b43aecc567f4-kube-api-access-hwg6j\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.854882 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv567\" (UniqueName: \"kubernetes.io/projected/8f14efb1-7152-4798-af6a-7a60d8cacccf-kube-api-access-cv567\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.854973 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855048 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f0420681-49bc-466a-976e-b43aecc567f4-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855123 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855214 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855303 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/45fe7b22-d9bd-490b-9a11-3a967263ee66-metrics-client-ca\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855380 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855455 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-textfile\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855519 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-tls\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855581 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8f14efb1-7152-4798-af6a-7a60d8cacccf-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855644 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/45fe7b22-d9bd-490b-9a11-3a967263ee66-root\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855716 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/8f14efb1-7152-4798-af6a-7a60d8cacccf-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855783 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/8f14efb1-7152-4798-af6a-7a60d8cacccf-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855861 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/45fe7b22-d9bd-490b-9a11-3a967263ee66-sys\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855926 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-wtmp\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.855998 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/f0420681-49bc-466a-976e-b43aecc567f4-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.857059 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/45fe7b22-d9bd-490b-9a11-3a967263ee66-root\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: E0929 17:13:54.857190 4667 secret.go:188] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: secret "openshift-state-metrics-tls" not found Sep 29 17:13:54 crc kubenswrapper[4667]: E0929 17:13:54.857249 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f14efb1-7152-4798-af6a-7a60d8cacccf-openshift-state-metrics-tls podName:8f14efb1-7152-4798-af6a-7a60d8cacccf nodeName:}" failed. No retries permitted until 2025-09-29 17:13:55.357233466 +0000 UTC m=+263.855080235 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/8f14efb1-7152-4798-af6a-7a60d8cacccf-openshift-state-metrics-tls") pod "openshift-state-metrics-566fddb674-qn4zj" (UID: "8f14efb1-7152-4798-af6a-7a60d8cacccf") : secret "openshift-state-metrics-tls" not found Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.857413 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/45fe7b22-d9bd-490b-9a11-3a967263ee66-sys\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.857498 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8f14efb1-7152-4798-af6a-7a60d8cacccf-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.857530 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-wtmp\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.857781 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-textfile\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.858256 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/45fe7b22-d9bd-490b-9a11-3a967263ee66-metrics-client-ca\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.862229 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-tls\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.862361 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/8f14efb1-7152-4798-af6a-7a60d8cacccf-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.863109 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/45fe7b22-d9bd-490b-9a11-3a967263ee66-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.866508 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbpkj\" (UniqueName: \"kubernetes.io/projected/45fe7b22-d9bd-490b-9a11-3a967263ee66-kube-api-access-kbpkj\") pod \"node-exporter-nxqpg\" (UID: \"45fe7b22-d9bd-490b-9a11-3a967263ee66\") " pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.869089 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv567\" (UniqueName: \"kubernetes.io/projected/8f14efb1-7152-4798-af6a-7a60d8cacccf-kube-api-access-cv567\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.957161 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/f0420681-49bc-466a-976e-b43aecc567f4-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.957221 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwg6j\" (UniqueName: \"kubernetes.io/projected/f0420681-49bc-466a-976e-b43aecc567f4-kube-api-access-hwg6j\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.957246 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.957264 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f0420681-49bc-466a-976e-b43aecc567f4-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.957282 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.957323 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.957741 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/f0420681-49bc-466a-976e-b43aecc567f4-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.958578 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f0420681-49bc-466a-976e-b43aecc567f4-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:54 crc kubenswrapper[4667]: I0929 17:13:54.971989 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwg6j\" (UniqueName: \"kubernetes.io/projected/f0420681-49bc-466a-976e-b43aecc567f4-kube-api-access-hwg6j\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.046833 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-nxqpg" Sep 29 17:13:55 crc kubenswrapper[4667]: W0929 17:13:55.062025 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45fe7b22_d9bd_490b_9a11_3a967263ee66.slice/crio-bf35815035b0d79871bfebed217416abbd96cf91417f3bee533b8aa0750629d6 WatchSource:0}: Error finding container bf35815035b0d79871bfebed217416abbd96cf91417f3bee533b8aa0750629d6: Status 404 returned error can't find the container with id bf35815035b0d79871bfebed217416abbd96cf91417f3bee533b8aa0750629d6 Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.361419 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/8f14efb1-7152-4798-af6a-7a60d8cacccf-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.363752 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/8f14efb1-7152-4798-af6a-7a60d8cacccf-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-qn4zj\" (UID: \"8f14efb1-7152-4798-af6a-7a60d8cacccf\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.636348 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.794133 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.795757 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.797897 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.798068 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.798429 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.798544 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.798565 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.798665 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-vx6d2" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.798731 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.799598 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.807193 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.848223 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.866797 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.866857 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/c9c697b4-73c9-48e6-a7b6-947ed8352095-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.866918 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c9c697b4-73c9-48e6-a7b6-947ed8352095-config-out\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.866954 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.866981 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.867213 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-config-volume\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.867255 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-web-config\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.867272 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mg4s\" (UniqueName: \"kubernetes.io/projected/c9c697b4-73c9-48e6-a7b6-947ed8352095-kube-api-access-4mg4s\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.867317 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c9c697b4-73c9-48e6-a7b6-947ed8352095-tls-assets\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.867348 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c9c697b4-73c9-48e6-a7b6-947ed8352095-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.867424 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9c697b4-73c9-48e6-a7b6-947ed8352095-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.867465 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.880376 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.888670 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:55 crc kubenswrapper[4667]: E0929 17:13:55.958665 4667 secret.go:188] Couldn't get secret openshift-monitoring/kube-state-metrics-tls: failed to sync secret cache: timed out waiting for the condition Sep 29 17:13:55 crc kubenswrapper[4667]: E0929 17:13:55.958687 4667 secret.go:188] Couldn't get secret openshift-monitoring/kube-state-metrics-kube-rbac-proxy-config: failed to sync secret cache: timed out waiting for the condition Sep 29 17:13:55 crc kubenswrapper[4667]: E0929 17:13:55.958725 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-tls podName:f0420681-49bc-466a-976e-b43aecc567f4 nodeName:}" failed. No retries permitted until 2025-09-29 17:13:56.45871136 +0000 UTC m=+264.956558129 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-tls" (UniqueName: "kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-tls") pod "kube-state-metrics-777cb5bd5d-jb9qv" (UID: "f0420681-49bc-466a-976e-b43aecc567f4") : failed to sync secret cache: timed out waiting for the condition Sep 29 17:13:55 crc kubenswrapper[4667]: E0929 17:13:55.958757 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-kube-rbac-proxy-config podName:f0420681-49bc-466a-976e-b43aecc567f4 nodeName:}" failed. No retries permitted until 2025-09-29 17:13:56.458741426 +0000 UTC m=+264.956588195 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-kube-rbac-proxy-config" (UniqueName: "kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-kube-rbac-proxy-config") pod "kube-state-metrics-777cb5bd5d-jb9qv" (UID: "f0420681-49bc-466a-976e-b43aecc567f4") : failed to sync secret cache: timed out waiting for the condition Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968096 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9c697b4-73c9-48e6-a7b6-947ed8352095-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968135 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968162 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968183 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/c9c697b4-73c9-48e6-a7b6-947ed8352095-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968221 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c9c697b4-73c9-48e6-a7b6-947ed8352095-config-out\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968244 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968260 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968286 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-config-volume\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968301 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-web-config\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968314 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mg4s\" (UniqueName: \"kubernetes.io/projected/c9c697b4-73c9-48e6-a7b6-947ed8352095-kube-api-access-4mg4s\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968337 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c9c697b4-73c9-48e6-a7b6-947ed8352095-tls-assets\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.968360 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c9c697b4-73c9-48e6-a7b6-947ed8352095-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.969384 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/c9c697b4-73c9-48e6-a7b6-947ed8352095-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.969621 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c9c697b4-73c9-48e6-a7b6-947ed8352095-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.969737 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9c697b4-73c9-48e6-a7b6-947ed8352095-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.971771 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.971835 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c9c697b4-73c9-48e6-a7b6-947ed8352095-config-out\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.972070 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-config-volume\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.972088 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.972345 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c9c697b4-73c9-48e6-a7b6-947ed8352095-tls-assets\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.972430 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-web-config\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.982465 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.982541 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/c9c697b4-73c9-48e6-a7b6-947ed8352095-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.984177 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mg4s\" (UniqueName: \"kubernetes.io/projected/c9c697b4-73c9-48e6-a7b6-947ed8352095-kube-api-access-4mg4s\") pod \"alertmanager-main-0\" (UID: \"c9c697b4-73c9-48e6-a7b6-947ed8352095\") " pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:55 crc kubenswrapper[4667]: I0929 17:13:55.998666 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.010896 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj"] Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.015983 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-nxqpg" event={"ID":"45fe7b22-d9bd-490b-9a11-3a967263ee66","Type":"ContainerStarted","Data":"bf35815035b0d79871bfebed217416abbd96cf91417f3bee533b8aa0750629d6"} Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.112870 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.301173 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.319353 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-9vm6b" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.473146 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.475043 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.475089 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:56 crc kubenswrapper[4667]: W0929 17:13:56.479497 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9c697b4_73c9_48e6_a7b6_947ed8352095.slice/crio-7624f9cf258b4dc6bb3b1585b8fc0283f0df4ece17cd46b24464d34764f4ac9e WatchSource:0}: Error finding container 7624f9cf258b4dc6bb3b1585b8fc0283f0df4ece17cd46b24464d34764f4ac9e: Status 404 returned error can't find the container with id 7624f9cf258b4dc6bb3b1585b8fc0283f0df4ece17cd46b24464d34764f4ac9e Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.479945 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.480213 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f0420681-49bc-466a-976e-b43aecc567f4-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-jb9qv\" (UID: \"f0420681-49bc-466a-976e-b43aecc567f4\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.558073 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.705056 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-797d5c77df-rpvdp"] Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.706443 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.707884 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.708205 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.708246 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-89ea5hknnd71m" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.708283 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-sfk58" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.708247 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.708358 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.711791 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.716730 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-797d5c77df-rpvdp"] Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.780665 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d2c52f95-0c97-4754-a304-e6ca2492e777-metrics-client-ca\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.780810 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.780959 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.781062 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-grpc-tls\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.781179 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-tls\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.781284 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.781365 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsgxh\" (UniqueName: \"kubernetes.io/projected/d2c52f95-0c97-4754-a304-e6ca2492e777-kube-api-access-vsgxh\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.781490 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.882329 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.882372 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-tls\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.882391 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsgxh\" (UniqueName: \"kubernetes.io/projected/d2c52f95-0c97-4754-a304-e6ca2492e777-kube-api-access-vsgxh\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.882417 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.882448 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.882465 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d2c52f95-0c97-4754-a304-e6ca2492e777-metrics-client-ca\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.882492 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.882514 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-grpc-tls\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.884450 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d2c52f95-0c97-4754-a304-e6ca2492e777-metrics-client-ca\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.886403 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-grpc-tls\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.886406 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.886501 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.887469 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-tls\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.887495 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.892405 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv"] Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.892989 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/d2c52f95-0c97-4754-a304-e6ca2492e777-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: I0929 17:13:56.896625 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsgxh\" (UniqueName: \"kubernetes.io/projected/d2c52f95-0c97-4754-a304-e6ca2492e777-kube-api-access-vsgxh\") pod \"thanos-querier-797d5c77df-rpvdp\" (UID: \"d2c52f95-0c97-4754-a304-e6ca2492e777\") " pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:56 crc kubenswrapper[4667]: W0929 17:13:56.899703 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0420681_49bc_466a_976e_b43aecc567f4.slice/crio-0c88fa3c96f6218fb1d54fc905122f395e8a75bef89e75310410d5ea05bc4dc6 WatchSource:0}: Error finding container 0c88fa3c96f6218fb1d54fc905122f395e8a75bef89e75310410d5ea05bc4dc6: Status 404 returned error can't find the container with id 0c88fa3c96f6218fb1d54fc905122f395e8a75bef89e75310410d5ea05bc4dc6 Sep 29 17:13:57 crc kubenswrapper[4667]: I0929 17:13:57.020604 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" event={"ID":"f0420681-49bc-466a-976e-b43aecc567f4","Type":"ContainerStarted","Data":"0c88fa3c96f6218fb1d54fc905122f395e8a75bef89e75310410d5ea05bc4dc6"} Sep 29 17:13:57 crc kubenswrapper[4667]: I0929 17:13:57.021512 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"c9c697b4-73c9-48e6-a7b6-947ed8352095","Type":"ContainerStarted","Data":"7624f9cf258b4dc6bb3b1585b8fc0283f0df4ece17cd46b24464d34764f4ac9e"} Sep 29 17:13:57 crc kubenswrapper[4667]: I0929 17:13:57.022982 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" event={"ID":"8f14efb1-7152-4798-af6a-7a60d8cacccf","Type":"ContainerStarted","Data":"57c7e576b7ba88a6404150fae8852ef827650cd69de13cdc971c621d6d638842"} Sep 29 17:13:57 crc kubenswrapper[4667]: I0929 17:13:57.023009 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" event={"ID":"8f14efb1-7152-4798-af6a-7a60d8cacccf","Type":"ContainerStarted","Data":"f6c06f520c2aab2354d1825feb6b512225120007e9588a27632e89a164331676"} Sep 29 17:13:57 crc kubenswrapper[4667]: I0929 17:13:57.023020 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" event={"ID":"8f14efb1-7152-4798-af6a-7a60d8cacccf","Type":"ContainerStarted","Data":"d06f31c965c3432588a9b660425a8aa81d1a8a91ebf6fd110d1ee19a6dd95e31"} Sep 29 17:13:57 crc kubenswrapper[4667]: I0929 17:13:57.023627 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:13:57 crc kubenswrapper[4667]: I0929 17:13:57.024188 4667 generic.go:334] "Generic (PLEG): container finished" podID="45fe7b22-d9bd-490b-9a11-3a967263ee66" containerID="0462f5a9f2f4d39879ee7e2d5337de5edcbe4bf66bdf6007e53d7c66d8c22325" exitCode=0 Sep 29 17:13:57 crc kubenswrapper[4667]: I0929 17:13:57.024588 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-nxqpg" event={"ID":"45fe7b22-d9bd-490b-9a11-3a967263ee66","Type":"ContainerDied","Data":"0462f5a9f2f4d39879ee7e2d5337de5edcbe4bf66bdf6007e53d7c66d8c22325"} Sep 29 17:13:57 crc kubenswrapper[4667]: I0929 17:13:57.367271 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-797d5c77df-rpvdp"] Sep 29 17:13:57 crc kubenswrapper[4667]: W0929 17:13:57.369126 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2c52f95_0c97_4754_a304_e6ca2492e777.slice/crio-829274d0ee30280e3a7022ae83a23795b04c3a6691a19ba0573a14454b9191b0 WatchSource:0}: Error finding container 829274d0ee30280e3a7022ae83a23795b04c3a6691a19ba0573a14454b9191b0: Status 404 returned error can't find the container with id 829274d0ee30280e3a7022ae83a23795b04c3a6691a19ba0573a14454b9191b0 Sep 29 17:13:58 crc kubenswrapper[4667]: I0929 17:13:58.031098 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-nxqpg" event={"ID":"45fe7b22-d9bd-490b-9a11-3a967263ee66","Type":"ContainerStarted","Data":"64f52aff43fcbdd5bb80e8a0fd983d11c756f72a6a648425e57cef1fe52c3d0c"} Sep 29 17:13:58 crc kubenswrapper[4667]: I0929 17:13:58.031357 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-nxqpg" event={"ID":"45fe7b22-d9bd-490b-9a11-3a967263ee66","Type":"ContainerStarted","Data":"5b7606686405774d9596bd3332d3ebf71247fc73892aedf0a37af0672cdc91cf"} Sep 29 17:13:58 crc kubenswrapper[4667]: I0929 17:13:58.032480 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" event={"ID":"8f14efb1-7152-4798-af6a-7a60d8cacccf","Type":"ContainerStarted","Data":"b102337c9f5484510adb16caed4701181e9b2eb796b7f5869f5bd67294c2c1cc"} Sep 29 17:13:58 crc kubenswrapper[4667]: I0929 17:13:58.033945 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" event={"ID":"d2c52f95-0c97-4754-a304-e6ca2492e777","Type":"ContainerStarted","Data":"829274d0ee30280e3a7022ae83a23795b04c3a6691a19ba0573a14454b9191b0"} Sep 29 17:13:58 crc kubenswrapper[4667]: I0929 17:13:58.043105 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-nxqpg" podStartSLOduration=2.689333804 podStartE2EDuration="4.043094519s" podCreationTimestamp="2025-09-29 17:13:54 +0000 UTC" firstStartedPulling="2025-09-29 17:13:55.06334695 +0000 UTC m=+263.561193719" lastFinishedPulling="2025-09-29 17:13:56.417107665 +0000 UTC m=+264.914954434" observedRunningTime="2025-09-29 17:13:58.040802184 +0000 UTC m=+266.538648953" watchObservedRunningTime="2025-09-29 17:13:58.043094519 +0000 UTC m=+266.540941288" Sep 29 17:13:58 crc kubenswrapper[4667]: I0929 17:13:58.053293 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-qn4zj" podStartSLOduration=2.366200802 podStartE2EDuration="4.05327744s" podCreationTimestamp="2025-09-29 17:13:54 +0000 UTC" firstStartedPulling="2025-09-29 17:13:56.198825665 +0000 UTC m=+264.696672424" lastFinishedPulling="2025-09-29 17:13:57.885902293 +0000 UTC m=+266.383749062" observedRunningTime="2025-09-29 17:13:58.05149835 +0000 UTC m=+266.549345119" watchObservedRunningTime="2025-09-29 17:13:58.05327744 +0000 UTC m=+266.551124210" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.040658 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" event={"ID":"f0420681-49bc-466a-976e-b43aecc567f4","Type":"ContainerStarted","Data":"f676671ff460d5c29b183df13b38a5fc8bd00a23bcca36be923f33ccf6380b24"} Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.044343 4667 generic.go:334] "Generic (PLEG): container finished" podID="c9c697b4-73c9-48e6-a7b6-947ed8352095" containerID="3d3d864ab7e56d016663fccdd8d2378d26b831f73986f13cb4aa99afe9130dfd" exitCode=0 Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.044540 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"c9c697b4-73c9-48e6-a7b6-947ed8352095","Type":"ContainerDied","Data":"3d3d864ab7e56d016663fccdd8d2378d26b831f73986f13cb4aa99afe9130dfd"} Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.542560 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-86d68f79b5-wkvqz"] Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.543212 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.559947 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86d68f79b5-wkvqz"] Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.615330 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-oauth-config\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.615402 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-oauth-serving-cert\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.615431 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-serving-cert\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.615456 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-trusted-ca-bundle\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.615489 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-console-config\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.615535 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-service-ca\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.615555 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kps9r\" (UniqueName: \"kubernetes.io/projected/71707e4c-df9e-463c-a704-d1973086442e-kube-api-access-kps9r\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.716483 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-service-ca\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.716529 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kps9r\" (UniqueName: \"kubernetes.io/projected/71707e4c-df9e-463c-a704-d1973086442e-kube-api-access-kps9r\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.716563 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-oauth-config\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.716606 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-oauth-serving-cert\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.716628 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-serving-cert\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.716651 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-trusted-ca-bundle\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.716684 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-console-config\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.717321 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-service-ca\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.717401 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-console-config\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.717559 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-oauth-serving-cert\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.718041 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-trusted-ca-bundle\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.721380 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-serving-cert\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.721460 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-oauth-config\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.729245 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kps9r\" (UniqueName: \"kubernetes.io/projected/71707e4c-df9e-463c-a704-d1973086442e-kube-api-access-kps9r\") pod \"console-86d68f79b5-wkvqz\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:13:59 crc kubenswrapper[4667]: I0929 17:13:59.854081 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.050345 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" event={"ID":"f0420681-49bc-466a-976e-b43aecc567f4","Type":"ContainerStarted","Data":"916b612a7d5a6fa6a747352afeffb773c19b3d12229dfb259a0664003cf8f3d1"} Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.050385 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" event={"ID":"f0420681-49bc-466a-976e-b43aecc567f4","Type":"ContainerStarted","Data":"f1de996949919d6962ffecc3c11eac830e42a3fa7ce37c92c8c6ae51d18f600c"} Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.063227 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-jb9qv" podStartSLOduration=4.095816141 podStartE2EDuration="6.063217131s" podCreationTimestamp="2025-09-29 17:13:54 +0000 UTC" firstStartedPulling="2025-09-29 17:13:56.901130069 +0000 UTC m=+265.398976837" lastFinishedPulling="2025-09-29 17:13:58.868531058 +0000 UTC m=+267.366377827" observedRunningTime="2025-09-29 17:14:00.062341874 +0000 UTC m=+268.560188643" watchObservedRunningTime="2025-09-29 17:14:00.063217131 +0000 UTC m=+268.561063901" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.139972 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-5847476974-qd4wh"] Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.140597 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.145413 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.145812 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.146016 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-l6wlm" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.146137 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.146241 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.147551 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-l6lq2qf57ip" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.147753 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-5847476974-qd4wh"] Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.221570 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/587a2614-fdf3-429c-a57c-5cde3ee0c577-metrics-server-audit-profiles\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.221612 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587a2614-fdf3-429c-a57c-5cde3ee0c577-client-ca-bundle\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.221638 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/587a2614-fdf3-429c-a57c-5cde3ee0c577-audit-log\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.221664 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxlrm\" (UniqueName: \"kubernetes.io/projected/587a2614-fdf3-429c-a57c-5cde3ee0c577-kube-api-access-xxlrm\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.221782 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/587a2614-fdf3-429c-a57c-5cde3ee0c577-secret-metrics-client-certs\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.221835 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/587a2614-fdf3-429c-a57c-5cde3ee0c577-secret-metrics-server-tls\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.221973 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/587a2614-fdf3-429c-a57c-5cde3ee0c577-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.322988 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/587a2614-fdf3-429c-a57c-5cde3ee0c577-metrics-server-audit-profiles\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.323226 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587a2614-fdf3-429c-a57c-5cde3ee0c577-client-ca-bundle\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.323262 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/587a2614-fdf3-429c-a57c-5cde3ee0c577-audit-log\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.323304 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxlrm\" (UniqueName: \"kubernetes.io/projected/587a2614-fdf3-429c-a57c-5cde3ee0c577-kube-api-access-xxlrm\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.323357 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/587a2614-fdf3-429c-a57c-5cde3ee0c577-secret-metrics-client-certs\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.323389 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/587a2614-fdf3-429c-a57c-5cde3ee0c577-secret-metrics-server-tls\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.323435 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/587a2614-fdf3-429c-a57c-5cde3ee0c577-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.324373 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/587a2614-fdf3-429c-a57c-5cde3ee0c577-audit-log\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.324401 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/587a2614-fdf3-429c-a57c-5cde3ee0c577-metrics-server-audit-profiles\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.324981 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/587a2614-fdf3-429c-a57c-5cde3ee0c577-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.326689 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/587a2614-fdf3-429c-a57c-5cde3ee0c577-client-ca-bundle\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.327045 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/587a2614-fdf3-429c-a57c-5cde3ee0c577-secret-metrics-server-tls\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.328413 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/587a2614-fdf3-429c-a57c-5cde3ee0c577-secret-metrics-client-certs\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.336097 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxlrm\" (UniqueName: \"kubernetes.io/projected/587a2614-fdf3-429c-a57c-5cde3ee0c577-kube-api-access-xxlrm\") pod \"metrics-server-5847476974-qd4wh\" (UID: \"587a2614-fdf3-429c-a57c-5cde3ee0c577\") " pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.455863 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.515392 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86d68f79b5-wkvqz"] Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.525503 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-cb6db7455-4hltx"] Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.526122 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.530065 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.530223 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Sep 29 17:14:00 crc kubenswrapper[4667]: W0929 17:14:00.534269 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71707e4c_df9e_463c_a704_d1973086442e.slice/crio-f0f88cf118de0c746ba0c380ab951e3ce3866e269d5735e0e9cd2ee5cd9cad49 WatchSource:0}: Error finding container f0f88cf118de0c746ba0c380ab951e3ce3866e269d5735e0e9cd2ee5cd9cad49: Status 404 returned error can't find the container with id f0f88cf118de0c746ba0c380ab951e3ce3866e269d5735e0e9cd2ee5cd9cad49 Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.534382 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-cb6db7455-4hltx"] Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.626496 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/a0f8aa79-3dbf-4890-a507-bf8a00198b85-monitoring-plugin-cert\") pod \"monitoring-plugin-cb6db7455-4hltx\" (UID: \"a0f8aa79-3dbf-4890-a507-bf8a00198b85\") " pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.727279 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/a0f8aa79-3dbf-4890-a507-bf8a00198b85-monitoring-plugin-cert\") pod \"monitoring-plugin-cb6db7455-4hltx\" (UID: \"a0f8aa79-3dbf-4890-a507-bf8a00198b85\") " pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.731514 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/a0f8aa79-3dbf-4890-a507-bf8a00198b85-monitoring-plugin-cert\") pod \"monitoring-plugin-cb6db7455-4hltx\" (UID: \"a0f8aa79-3dbf-4890-a507-bf8a00198b85\") " pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.827060 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-5847476974-qd4wh"] Sep 29 17:14:00 crc kubenswrapper[4667]: W0929 17:14:00.831878 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod587a2614_fdf3_429c_a57c_5cde3ee0c577.slice/crio-4a50515e7befe7611f3f97c0a3691d4c033876a1ca0545db0f9301c5092bb182 WatchSource:0}: Error finding container 4a50515e7befe7611f3f97c0a3691d4c033876a1ca0545db0f9301c5092bb182: Status 404 returned error can't find the container with id 4a50515e7befe7611f3f97c0a3691d4c033876a1ca0545db0f9301c5092bb182 Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.851404 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.905648 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.907379 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.909536 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.909799 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.909939 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-7sqmd59g60a9p" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.910928 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.911158 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.911245 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.913704 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.914766 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.914829 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.914931 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.915133 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.916542 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-9g8bz" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.920998 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.924109 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936536 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwtxs\" (UniqueName: \"kubernetes.io/projected/d7cd8150-68d6-47a5-9fdd-9e6186792497-kube-api-access-mwtxs\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936573 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936610 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936627 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936679 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936717 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936734 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/d7cd8150-68d6-47a5-9fdd-9e6186792497-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936751 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d7cd8150-68d6-47a5-9fdd-9e6186792497-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936764 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936795 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936809 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936831 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936874 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-config\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936889 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936906 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-web-config\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936922 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d7cd8150-68d6-47a5-9fdd-9e6186792497-config-out\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936944 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:00 crc kubenswrapper[4667]: I0929 17:14:00.936961 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037715 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037746 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037786 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037810 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037824 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/d7cd8150-68d6-47a5-9fdd-9e6186792497-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037857 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d7cd8150-68d6-47a5-9fdd-9e6186792497-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037871 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037890 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037904 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037922 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037939 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-config\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037951 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037968 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-web-config\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037983 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d7cd8150-68d6-47a5-9fdd-9e6186792497-config-out\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.037999 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.038015 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.038039 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwtxs\" (UniqueName: \"kubernetes.io/projected/d7cd8150-68d6-47a5-9fdd-9e6186792497-kube-api-access-mwtxs\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.038055 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.038563 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/d7cd8150-68d6-47a5-9fdd-9e6186792497-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.039234 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.040133 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.040503 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.041852 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.041991 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.042119 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.042912 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.043072 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d7cd8150-68d6-47a5-9fdd-9e6186792497-config-out\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.043665 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-config\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.044102 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.044147 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d7cd8150-68d6-47a5-9fdd-9e6186792497-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.044156 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.045152 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.047001 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.051518 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/d7cd8150-68d6-47a5-9fdd-9e6186792497-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.052877 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwtxs\" (UniqueName: \"kubernetes.io/projected/d7cd8150-68d6-47a5-9fdd-9e6186792497-kube-api-access-mwtxs\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.057818 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d7cd8150-68d6-47a5-9fdd-9e6186792497-web-config\") pod \"prometheus-k8s-0\" (UID: \"d7cd8150-68d6-47a5-9fdd-9e6186792497\") " pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.058251 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-5847476974-qd4wh" event={"ID":"587a2614-fdf3-429c-a57c-5cde3ee0c577","Type":"ContainerStarted","Data":"4a50515e7befe7611f3f97c0a3691d4c033876a1ca0545db0f9301c5092bb182"} Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.060392 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" event={"ID":"d2c52f95-0c97-4754-a304-e6ca2492e777","Type":"ContainerStarted","Data":"7fbfedaf0ef8cac8dbd80d73291d12753fbd0a73c62f6dde1891e22a331811fd"} Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.060433 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" event={"ID":"d2c52f95-0c97-4754-a304-e6ca2492e777","Type":"ContainerStarted","Data":"2c1113aed476929ca5e5e114dff56586e3235b0cfe46372792e257ac827cf653"} Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.060464 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" event={"ID":"d2c52f95-0c97-4754-a304-e6ca2492e777","Type":"ContainerStarted","Data":"d0d2048adc33e26bf9c92295bd4ddc37fcfd12a1968707f21098f70c7e82ec45"} Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.066146 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86d68f79b5-wkvqz" event={"ID":"71707e4c-df9e-463c-a704-d1973086442e","Type":"ContainerStarted","Data":"07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df"} Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.066182 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86d68f79b5-wkvqz" event={"ID":"71707e4c-df9e-463c-a704-d1973086442e","Type":"ContainerStarted","Data":"f0f88cf118de0c746ba0c380ab951e3ce3866e269d5735e0e9cd2ee5cd9cad49"} Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.223214 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.240219 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-86d68f79b5-wkvqz" podStartSLOduration=2.240205597 podStartE2EDuration="2.240205597s" podCreationTimestamp="2025-09-29 17:13:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:14:01.080153818 +0000 UTC m=+269.578000588" watchObservedRunningTime="2025-09-29 17:14:01.240205597 +0000 UTC m=+269.738052355" Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.242776 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-cb6db7455-4hltx"] Sep 29 17:14:01 crc kubenswrapper[4667]: W0929 17:14:01.251696 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0f8aa79_3dbf_4890_a507_bf8a00198b85.slice/crio-6f012767150543c98734ba8d51a38346bd39580d9f7751d18cd93725d9d927ee WatchSource:0}: Error finding container 6f012767150543c98734ba8d51a38346bd39580d9f7751d18cd93725d9d927ee: Status 404 returned error can't find the container with id 6f012767150543c98734ba8d51a38346bd39580d9f7751d18cd93725d9d927ee Sep 29 17:14:01 crc kubenswrapper[4667]: I0929 17:14:01.580538 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Sep 29 17:14:01 crc kubenswrapper[4667]: W0929 17:14:01.584952 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7cd8150_68d6_47a5_9fdd_9e6186792497.slice/crio-8b087a70792508acd14f62bb1d22dd95854971d3d63da16b554557ce47210f73 WatchSource:0}: Error finding container 8b087a70792508acd14f62bb1d22dd95854971d3d63da16b554557ce47210f73: Status 404 returned error can't find the container with id 8b087a70792508acd14f62bb1d22dd95854971d3d63da16b554557ce47210f73 Sep 29 17:14:02 crc kubenswrapper[4667]: I0929 17:14:02.071505 4667 generic.go:334] "Generic (PLEG): container finished" podID="d7cd8150-68d6-47a5-9fdd-9e6186792497" containerID="66fd3d846db80e3deee35aaad2c562cedb5b0d3f1acaa5dcb34f0977ca2f0349" exitCode=0 Sep 29 17:14:02 crc kubenswrapper[4667]: I0929 17:14:02.071544 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"d7cd8150-68d6-47a5-9fdd-9e6186792497","Type":"ContainerDied","Data":"66fd3d846db80e3deee35aaad2c562cedb5b0d3f1acaa5dcb34f0977ca2f0349"} Sep 29 17:14:02 crc kubenswrapper[4667]: I0929 17:14:02.071587 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"d7cd8150-68d6-47a5-9fdd-9e6186792497","Type":"ContainerStarted","Data":"8b087a70792508acd14f62bb1d22dd95854971d3d63da16b554557ce47210f73"} Sep 29 17:14:02 crc kubenswrapper[4667]: I0929 17:14:02.075058 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" event={"ID":"a0f8aa79-3dbf-4890-a507-bf8a00198b85","Type":"ContainerStarted","Data":"6f012767150543c98734ba8d51a38346bd39580d9f7751d18cd93725d9d927ee"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.082213 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"c9c697b4-73c9-48e6-a7b6-947ed8352095","Type":"ContainerStarted","Data":"c9526c7a57f2af66dd6dc603f4e43f5a6ffa157e984f35169c0f0c970ba4cc8e"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.082464 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"c9c697b4-73c9-48e6-a7b6-947ed8352095","Type":"ContainerStarted","Data":"8d85b9d571bcbbad0851ececf8d2e3817927e5a856ca9aadcb6abd9f64838813"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.082476 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"c9c697b4-73c9-48e6-a7b6-947ed8352095","Type":"ContainerStarted","Data":"d6f1077338e26006a476d6ac48f605891af6d33b157f2c7cb7fea1fd95373933"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.082484 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"c9c697b4-73c9-48e6-a7b6-947ed8352095","Type":"ContainerStarted","Data":"29d5f7ae75bf7a8bcd1616381095cc79826ab3d2f41b5f2ed0d3e1dc8fee7cf0"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.082491 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"c9c697b4-73c9-48e6-a7b6-947ed8352095","Type":"ContainerStarted","Data":"339d9475099a0ea69e55fe89145b66cdaf4d640a02e6fb752c5fd4e54affd47c"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.082499 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"c9c697b4-73c9-48e6-a7b6-947ed8352095","Type":"ContainerStarted","Data":"8d1683c6ba9a0c813cf2145bcdac623ca4275eb897a85b29078c78e1b17b0cdd"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.085866 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" event={"ID":"d2c52f95-0c97-4754-a304-e6ca2492e777","Type":"ContainerStarted","Data":"c1d5c52e38c3a2981c7fa22ce91a8b6dacd964103c6412599806869beefa1ee2"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.085891 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" event={"ID":"d2c52f95-0c97-4754-a304-e6ca2492e777","Type":"ContainerStarted","Data":"71bbdee772cd17250825a274a9b240d073559d47a2c8f043780807bfd53371e4"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.085901 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" event={"ID":"d2c52f95-0c97-4754-a304-e6ca2492e777","Type":"ContainerStarted","Data":"f760a1e7033481a5d8bd3a57a7af673dccdefc7ce5e8c9b4d41a53508d72863e"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.086337 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.087390 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-5847476974-qd4wh" event={"ID":"587a2614-fdf3-429c-a57c-5cde3ee0c577","Type":"ContainerStarted","Data":"c16b33bfe537d9c19b3106a32880d6c9bfdd650967856aaa46a0d2e36289d9a0"} Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.105820 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.475088995 podStartE2EDuration="8.105810401s" podCreationTimestamp="2025-09-29 17:13:55 +0000 UTC" firstStartedPulling="2025-09-29 17:13:56.480699323 +0000 UTC m=+264.978546092" lastFinishedPulling="2025-09-29 17:14:02.111420729 +0000 UTC m=+270.609267498" observedRunningTime="2025-09-29 17:14:03.098998301 +0000 UTC m=+271.596845070" watchObservedRunningTime="2025-09-29 17:14:03.105810401 +0000 UTC m=+271.603657160" Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.127369 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" podStartSLOduration=2.391567792 podStartE2EDuration="7.127351657s" podCreationTimestamp="2025-09-29 17:13:56 +0000 UTC" firstStartedPulling="2025-09-29 17:13:57.372744009 +0000 UTC m=+265.870590777" lastFinishedPulling="2025-09-29 17:14:02.108527873 +0000 UTC m=+270.606374642" observedRunningTime="2025-09-29 17:14:03.126192555 +0000 UTC m=+271.624039324" watchObservedRunningTime="2025-09-29 17:14:03.127351657 +0000 UTC m=+271.625198427" Sep 29 17:14:03 crc kubenswrapper[4667]: I0929 17:14:03.138937 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-5847476974-qd4wh" podStartSLOduration=1.456678139 podStartE2EDuration="3.138922159s" podCreationTimestamp="2025-09-29 17:14:00 +0000 UTC" firstStartedPulling="2025-09-29 17:14:00.833778849 +0000 UTC m=+269.331625618" lastFinishedPulling="2025-09-29 17:14:02.516022869 +0000 UTC m=+271.013869638" observedRunningTime="2025-09-29 17:14:03.136697816 +0000 UTC m=+271.634544605" watchObservedRunningTime="2025-09-29 17:14:03.138922159 +0000 UTC m=+271.636768928" Sep 29 17:14:04 crc kubenswrapper[4667]: I0929 17:14:04.092958 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" event={"ID":"a0f8aa79-3dbf-4890-a507-bf8a00198b85","Type":"ContainerStarted","Data":"fdc14763c2fd628a389cede262388a261d40bbf0bac946f4f50ee6ee9444a017"} Sep 29 17:14:04 crc kubenswrapper[4667]: I0929 17:14:04.103795 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" podStartSLOduration=1.57238291 podStartE2EDuration="4.103782611s" podCreationTimestamp="2025-09-29 17:14:00 +0000 UTC" firstStartedPulling="2025-09-29 17:14:01.253496367 +0000 UTC m=+269.751343136" lastFinishedPulling="2025-09-29 17:14:03.784896067 +0000 UTC m=+272.282742837" observedRunningTime="2025-09-29 17:14:04.101939336 +0000 UTC m=+272.599786106" watchObservedRunningTime="2025-09-29 17:14:04.103782611 +0000 UTC m=+272.601629380" Sep 29 17:14:04 crc kubenswrapper[4667]: I0929 17:14:04.623866 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-xmbzg" Sep 29 17:14:04 crc kubenswrapper[4667]: I0929 17:14:04.659551 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f44b9"] Sep 29 17:14:05 crc kubenswrapper[4667]: I0929 17:14:05.102632 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"d7cd8150-68d6-47a5-9fdd-9e6186792497","Type":"ContainerStarted","Data":"818df9e0c38602e3040cc17fce713f791c58921443f03440e85c7577c28ce0aa"} Sep 29 17:14:05 crc kubenswrapper[4667]: I0929 17:14:05.102871 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"d7cd8150-68d6-47a5-9fdd-9e6186792497","Type":"ContainerStarted","Data":"403721dcf107e1d99a58c2fa0cbc5a2faaae0b91ab6b6dcb7beb2c9ccbbf1d1c"} Sep 29 17:14:05 crc kubenswrapper[4667]: I0929 17:14:05.102883 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"d7cd8150-68d6-47a5-9fdd-9e6186792497","Type":"ContainerStarted","Data":"967ecc619471e84c7ad90a74489bf44c2aa862b7e12b2ba89ca0beb53f0b0067"} Sep 29 17:14:05 crc kubenswrapper[4667]: I0929 17:14:05.102895 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" Sep 29 17:14:05 crc kubenswrapper[4667]: I0929 17:14:05.102905 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"d7cd8150-68d6-47a5-9fdd-9e6186792497","Type":"ContainerStarted","Data":"5f61ab543fb8f913c94137497b6446cfe06d0b15cd4b03508c7b4bbfc3fb2e6e"} Sep 29 17:14:05 crc kubenswrapper[4667]: I0929 17:14:05.107287 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-cb6db7455-4hltx" Sep 29 17:14:05 crc kubenswrapper[4667]: I0929 17:14:05.110045 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-797d5c77df-rpvdp" Sep 29 17:14:06 crc kubenswrapper[4667]: I0929 17:14:06.111370 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"d7cd8150-68d6-47a5-9fdd-9e6186792497","Type":"ContainerStarted","Data":"e74133f89fba7de3d99b86aa84674a6643d3243ebf77cf8094ec1f4c04a629db"} Sep 29 17:14:06 crc kubenswrapper[4667]: I0929 17:14:06.111405 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"d7cd8150-68d6-47a5-9fdd-9e6186792497","Type":"ContainerStarted","Data":"2d613343bc8af257617565c4ce72d1022d53c4bbca3657da78756d4f23c539b8"} Sep 29 17:14:06 crc kubenswrapper[4667]: I0929 17:14:06.132761 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=3.5223378739999998 podStartE2EDuration="6.132749971s" podCreationTimestamp="2025-09-29 17:14:00 +0000 UTC" firstStartedPulling="2025-09-29 17:14:02.072895741 +0000 UTC m=+270.570742511" lastFinishedPulling="2025-09-29 17:14:04.683307839 +0000 UTC m=+273.181154608" observedRunningTime="2025-09-29 17:14:06.130686714 +0000 UTC m=+274.628533484" watchObservedRunningTime="2025-09-29 17:14:06.132749971 +0000 UTC m=+274.630596739" Sep 29 17:14:06 crc kubenswrapper[4667]: I0929 17:14:06.224023 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:14:09 crc kubenswrapper[4667]: I0929 17:14:09.854320 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:14:09 crc kubenswrapper[4667]: I0929 17:14:09.854369 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:14:09 crc kubenswrapper[4667]: I0929 17:14:09.858160 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:14:10 crc kubenswrapper[4667]: I0929 17:14:10.135333 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:14:10 crc kubenswrapper[4667]: I0929 17:14:10.164851 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-qhc55"] Sep 29 17:14:20 crc kubenswrapper[4667]: I0929 17:14:20.456062 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:20 crc kubenswrapper[4667]: I0929 17:14:20.456415 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:29 crc kubenswrapper[4667]: I0929 17:14:29.686676 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" podUID="f2261944-710e-493f-9360-de4dea650ae5" containerName="registry" containerID="cri-o://b263e364ded51d0ec1708e37a2240252abf3690d0519b50caac9fc1eb1b58bdb" gracePeriod=30 Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.215737 4667 generic.go:334] "Generic (PLEG): container finished" podID="f2261944-710e-493f-9360-de4dea650ae5" containerID="b263e364ded51d0ec1708e37a2240252abf3690d0519b50caac9fc1eb1b58bdb" exitCode=0 Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.215826 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" event={"ID":"f2261944-710e-493f-9360-de4dea650ae5","Type":"ContainerDied","Data":"b263e364ded51d0ec1708e37a2240252abf3690d0519b50caac9fc1eb1b58bdb"} Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.216052 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" event={"ID":"f2261944-710e-493f-9360-de4dea650ae5","Type":"ContainerDied","Data":"9a868f8cda24244d43f0996bd5b9cd61871f1c34b168efa5688d7ed8ff520fb9"} Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.216066 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a868f8cda24244d43f0996bd5b9cd61871f1c34b168efa5688d7ed8ff520fb9" Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.941052 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.991886 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-trusted-ca\") pod \"f2261944-710e-493f-9360-de4dea650ae5\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.991996 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f2261944-710e-493f-9360-de4dea650ae5-installation-pull-secrets\") pod \"f2261944-710e-493f-9360-de4dea650ae5\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.992047 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-bound-sa-token\") pod \"f2261944-710e-493f-9360-de4dea650ae5\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.992111 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw7n4\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-kube-api-access-cw7n4\") pod \"f2261944-710e-493f-9360-de4dea650ae5\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.992720 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f2261944-710e-493f-9360-de4dea650ae5-ca-trust-extracted\") pod \"f2261944-710e-493f-9360-de4dea650ae5\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.992767 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-registry-tls\") pod \"f2261944-710e-493f-9360-de4dea650ae5\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.992831 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-registry-certificates\") pod \"f2261944-710e-493f-9360-de4dea650ae5\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.992950 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"f2261944-710e-493f-9360-de4dea650ae5\" (UID: \"f2261944-710e-493f-9360-de4dea650ae5\") " Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.993398 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "f2261944-710e-493f-9360-de4dea650ae5" (UID: "f2261944-710e-493f-9360-de4dea650ae5"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.993308 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "f2261944-710e-493f-9360-de4dea650ae5" (UID: "f2261944-710e-493f-9360-de4dea650ae5"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.996185 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2261944-710e-493f-9360-de4dea650ae5-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "f2261944-710e-493f-9360-de4dea650ae5" (UID: "f2261944-710e-493f-9360-de4dea650ae5"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.996321 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "f2261944-710e-493f-9360-de4dea650ae5" (UID: "f2261944-710e-493f-9360-de4dea650ae5"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.996651 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "f2261944-710e-493f-9360-de4dea650ae5" (UID: "f2261944-710e-493f-9360-de4dea650ae5"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:14:30 crc kubenswrapper[4667]: I0929 17:14:30.998814 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-kube-api-access-cw7n4" (OuterVolumeSpecName: "kube-api-access-cw7n4") pod "f2261944-710e-493f-9360-de4dea650ae5" (UID: "f2261944-710e-493f-9360-de4dea650ae5"). InnerVolumeSpecName "kube-api-access-cw7n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.001578 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "f2261944-710e-493f-9360-de4dea650ae5" (UID: "f2261944-710e-493f-9360-de4dea650ae5"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.005001 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2261944-710e-493f-9360-de4dea650ae5-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "f2261944-710e-493f-9360-de4dea650ae5" (UID: "f2261944-710e-493f-9360-de4dea650ae5"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.094699 4667 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-bound-sa-token\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.094724 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw7n4\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-kube-api-access-cw7n4\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.094735 4667 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f2261944-710e-493f-9360-de4dea650ae5-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.094745 4667 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f2261944-710e-493f-9360-de4dea650ae5-registry-tls\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.094753 4667 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-registry-certificates\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.094760 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2261944-710e-493f-9360-de4dea650ae5-trusted-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.094769 4667 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f2261944-710e-493f-9360-de4dea650ae5-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.220084 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-f44b9" Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.238307 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f44b9"] Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.242219 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-f44b9"] Sep 29 17:14:31 crc kubenswrapper[4667]: I0929 17:14:31.821384 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2261944-710e-493f-9360-de4dea650ae5" path="/var/lib/kubelet/pods/f2261944-710e-493f-9360-de4dea650ae5/volumes" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.189222 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-qhc55" podUID="cab548cd-ee10-421c-9648-02dd9ad58dfa" containerName="console" containerID="cri-o://8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80" gracePeriod=15 Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.465491 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-qhc55_cab548cd-ee10-421c-9648-02dd9ad58dfa/console/0.log" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.465701 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.536345 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-trusted-ca-bundle\") pod \"cab548cd-ee10-421c-9648-02dd9ad58dfa\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.536405 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-serving-cert\") pod \"cab548cd-ee10-421c-9648-02dd9ad58dfa\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.536446 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6sn9\" (UniqueName: \"kubernetes.io/projected/cab548cd-ee10-421c-9648-02dd9ad58dfa-kube-api-access-q6sn9\") pod \"cab548cd-ee10-421c-9648-02dd9ad58dfa\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.536469 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-config\") pod \"cab548cd-ee10-421c-9648-02dd9ad58dfa\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.536492 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-oauth-serving-cert\") pod \"cab548cd-ee10-421c-9648-02dd9ad58dfa\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.536512 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-oauth-config\") pod \"cab548cd-ee10-421c-9648-02dd9ad58dfa\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.536527 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-service-ca\") pod \"cab548cd-ee10-421c-9648-02dd9ad58dfa\" (UID: \"cab548cd-ee10-421c-9648-02dd9ad58dfa\") " Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.537066 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "cab548cd-ee10-421c-9648-02dd9ad58dfa" (UID: "cab548cd-ee10-421c-9648-02dd9ad58dfa"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.537078 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-config" (OuterVolumeSpecName: "console-config") pod "cab548cd-ee10-421c-9648-02dd9ad58dfa" (UID: "cab548cd-ee10-421c-9648-02dd9ad58dfa"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.537086 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-service-ca" (OuterVolumeSpecName: "service-ca") pod "cab548cd-ee10-421c-9648-02dd9ad58dfa" (UID: "cab548cd-ee10-421c-9648-02dd9ad58dfa"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.537564 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "cab548cd-ee10-421c-9648-02dd9ad58dfa" (UID: "cab548cd-ee10-421c-9648-02dd9ad58dfa"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.540683 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "cab548cd-ee10-421c-9648-02dd9ad58dfa" (UID: "cab548cd-ee10-421c-9648-02dd9ad58dfa"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.541061 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cab548cd-ee10-421c-9648-02dd9ad58dfa-kube-api-access-q6sn9" (OuterVolumeSpecName: "kube-api-access-q6sn9") pod "cab548cd-ee10-421c-9648-02dd9ad58dfa" (UID: "cab548cd-ee10-421c-9648-02dd9ad58dfa"). InnerVolumeSpecName "kube-api-access-q6sn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.541071 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "cab548cd-ee10-421c-9648-02dd9ad58dfa" (UID: "cab548cd-ee10-421c-9648-02dd9ad58dfa"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.637904 4667 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.637933 4667 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.637943 4667 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.637952 4667 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.637959 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cab548cd-ee10-421c-9648-02dd9ad58dfa-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.637968 4667 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cab548cd-ee10-421c-9648-02dd9ad58dfa-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:35 crc kubenswrapper[4667]: I0929 17:14:35.637976 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6sn9\" (UniqueName: \"kubernetes.io/projected/cab548cd-ee10-421c-9648-02dd9ad58dfa-kube-api-access-q6sn9\") on node \"crc\" DevicePath \"\"" Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.242615 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-qhc55_cab548cd-ee10-421c-9648-02dd9ad58dfa/console/0.log" Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.242837 4667 generic.go:334] "Generic (PLEG): container finished" podID="cab548cd-ee10-421c-9648-02dd9ad58dfa" containerID="8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80" exitCode=2 Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.242879 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qhc55" event={"ID":"cab548cd-ee10-421c-9648-02dd9ad58dfa","Type":"ContainerDied","Data":"8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80"} Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.242907 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qhc55" event={"ID":"cab548cd-ee10-421c-9648-02dd9ad58dfa","Type":"ContainerDied","Data":"f369ea26b675f56bf991251edd9e5e94e08698c54213c660939d543c07c7adbc"} Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.242923 4667 scope.go:117] "RemoveContainer" containerID="8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80" Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.243059 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qhc55" Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.254350 4667 scope.go:117] "RemoveContainer" containerID="8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80" Sep 29 17:14:36 crc kubenswrapper[4667]: E0929 17:14:36.254872 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80\": container with ID starting with 8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80 not found: ID does not exist" containerID="8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80" Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.254907 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80"} err="failed to get container status \"8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80\": rpc error: code = NotFound desc = could not find container \"8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80\": container with ID starting with 8bfda0d571f03ff5a6a0e3c673acd91611afda60097669bd84ce426ba5217d80 not found: ID does not exist" Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.257713 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-qhc55"] Sep 29 17:14:36 crc kubenswrapper[4667]: I0929 17:14:36.259933 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-qhc55"] Sep 29 17:14:37 crc kubenswrapper[4667]: I0929 17:14:37.820512 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cab548cd-ee10-421c-9648-02dd9ad58dfa" path="/var/lib/kubelet/pods/cab548cd-ee10-421c-9648-02dd9ad58dfa/volumes" Sep 29 17:14:40 crc kubenswrapper[4667]: I0929 17:14:40.461160 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:14:40 crc kubenswrapper[4667]: I0929 17:14:40.464262 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-5847476974-qd4wh" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.129762 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm"] Sep 29 17:15:00 crc kubenswrapper[4667]: E0929 17:15:00.130403 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2261944-710e-493f-9360-de4dea650ae5" containerName="registry" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.130417 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2261944-710e-493f-9360-de4dea650ae5" containerName="registry" Sep 29 17:15:00 crc kubenswrapper[4667]: E0929 17:15:00.130432 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab548cd-ee10-421c-9648-02dd9ad58dfa" containerName="console" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.130445 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab548cd-ee10-421c-9648-02dd9ad58dfa" containerName="console" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.130707 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="cab548cd-ee10-421c-9648-02dd9ad58dfa" containerName="console" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.130759 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2261944-710e-493f-9360-de4dea650ae5" containerName="registry" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.131638 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.132775 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.133146 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.136499 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm"] Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.188390 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c90c115b-e49a-4555-b0e0-f0ef85847110-config-volume\") pod \"collect-profiles-29319435-tdznm\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.188445 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rghlv\" (UniqueName: \"kubernetes.io/projected/c90c115b-e49a-4555-b0e0-f0ef85847110-kube-api-access-rghlv\") pod \"collect-profiles-29319435-tdznm\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.188484 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c90c115b-e49a-4555-b0e0-f0ef85847110-secret-volume\") pod \"collect-profiles-29319435-tdznm\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.289386 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rghlv\" (UniqueName: \"kubernetes.io/projected/c90c115b-e49a-4555-b0e0-f0ef85847110-kube-api-access-rghlv\") pod \"collect-profiles-29319435-tdznm\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.289435 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c90c115b-e49a-4555-b0e0-f0ef85847110-secret-volume\") pod \"collect-profiles-29319435-tdznm\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.289483 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c90c115b-e49a-4555-b0e0-f0ef85847110-config-volume\") pod \"collect-profiles-29319435-tdznm\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.290151 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c90c115b-e49a-4555-b0e0-f0ef85847110-config-volume\") pod \"collect-profiles-29319435-tdznm\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.293509 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c90c115b-e49a-4555-b0e0-f0ef85847110-secret-volume\") pod \"collect-profiles-29319435-tdznm\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.302916 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rghlv\" (UniqueName: \"kubernetes.io/projected/c90c115b-e49a-4555-b0e0-f0ef85847110-kube-api-access-rghlv\") pod \"collect-profiles-29319435-tdznm\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.446193 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:00 crc kubenswrapper[4667]: I0929 17:15:00.780521 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm"] Sep 29 17:15:01 crc kubenswrapper[4667]: I0929 17:15:01.224020 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:15:01 crc kubenswrapper[4667]: I0929 17:15:01.245292 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:15:01 crc kubenswrapper[4667]: I0929 17:15:01.352473 4667 generic.go:334] "Generic (PLEG): container finished" podID="c90c115b-e49a-4555-b0e0-f0ef85847110" containerID="bab673b5087eca42c32d59ee75551b51feff424ef65aba52181595b39fa9f3b1" exitCode=0 Sep 29 17:15:01 crc kubenswrapper[4667]: I0929 17:15:01.352521 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" event={"ID":"c90c115b-e49a-4555-b0e0-f0ef85847110","Type":"ContainerDied","Data":"bab673b5087eca42c32d59ee75551b51feff424ef65aba52181595b39fa9f3b1"} Sep 29 17:15:01 crc kubenswrapper[4667]: I0929 17:15:01.352547 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" event={"ID":"c90c115b-e49a-4555-b0e0-f0ef85847110","Type":"ContainerStarted","Data":"7a13b559c163c4dd8df069f5d9648e995e63aaf092a8a2ae89083681e0073726"} Sep 29 17:15:01 crc kubenswrapper[4667]: I0929 17:15:01.375648 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.523916 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.621826 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c90c115b-e49a-4555-b0e0-f0ef85847110-secret-volume\") pod \"c90c115b-e49a-4555-b0e0-f0ef85847110\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.621882 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c90c115b-e49a-4555-b0e0-f0ef85847110-config-volume\") pod \"c90c115b-e49a-4555-b0e0-f0ef85847110\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.621980 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rghlv\" (UniqueName: \"kubernetes.io/projected/c90c115b-e49a-4555-b0e0-f0ef85847110-kube-api-access-rghlv\") pod \"c90c115b-e49a-4555-b0e0-f0ef85847110\" (UID: \"c90c115b-e49a-4555-b0e0-f0ef85847110\") " Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.624425 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c90c115b-e49a-4555-b0e0-f0ef85847110-config-volume" (OuterVolumeSpecName: "config-volume") pod "c90c115b-e49a-4555-b0e0-f0ef85847110" (UID: "c90c115b-e49a-4555-b0e0-f0ef85847110"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.627218 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c90c115b-e49a-4555-b0e0-f0ef85847110-kube-api-access-rghlv" (OuterVolumeSpecName: "kube-api-access-rghlv") pod "c90c115b-e49a-4555-b0e0-f0ef85847110" (UID: "c90c115b-e49a-4555-b0e0-f0ef85847110"). InnerVolumeSpecName "kube-api-access-rghlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.627481 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c90c115b-e49a-4555-b0e0-f0ef85847110-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c90c115b-e49a-4555-b0e0-f0ef85847110" (UID: "c90c115b-e49a-4555-b0e0-f0ef85847110"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.723338 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rghlv\" (UniqueName: \"kubernetes.io/projected/c90c115b-e49a-4555-b0e0-f0ef85847110-kube-api-access-rghlv\") on node \"crc\" DevicePath \"\"" Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.723367 4667 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c90c115b-e49a-4555-b0e0-f0ef85847110-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 17:15:02 crc kubenswrapper[4667]: I0929 17:15:02.723378 4667 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c90c115b-e49a-4555-b0e0-f0ef85847110-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 17:15:03 crc kubenswrapper[4667]: I0929 17:15:03.361555 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" event={"ID":"c90c115b-e49a-4555-b0e0-f0ef85847110","Type":"ContainerDied","Data":"7a13b559c163c4dd8df069f5d9648e995e63aaf092a8a2ae89083681e0073726"} Sep 29 17:15:03 crc kubenswrapper[4667]: I0929 17:15:03.361592 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a13b559c163c4dd8df069f5d9648e995e63aaf092a8a2ae89083681e0073726" Sep 29 17:15:03 crc kubenswrapper[4667]: I0929 17:15:03.361655 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319435-tdznm" Sep 29 17:15:26 crc kubenswrapper[4667]: I0929 17:15:26.897476 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:15:26 crc kubenswrapper[4667]: I0929 17:15:26.897808 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.164547 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-785bd4864d-4f7hk"] Sep 29 17:15:50 crc kubenswrapper[4667]: E0929 17:15:50.165103 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90c115b-e49a-4555-b0e0-f0ef85847110" containerName="collect-profiles" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.165116 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90c115b-e49a-4555-b0e0-f0ef85847110" containerName="collect-profiles" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.165208 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="c90c115b-e49a-4555-b0e0-f0ef85847110" containerName="collect-profiles" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.165558 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.175990 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-785bd4864d-4f7hk"] Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.293474 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-trusted-ca-bundle\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.293549 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-serving-cert\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.293591 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-oauth-serving-cert\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.293661 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-oauth-config\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.293697 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nthfk\" (UniqueName: \"kubernetes.io/projected/8b280c9f-d84d-4cda-a63a-800f116d0cca-kube-api-access-nthfk\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.293723 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-service-ca\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.293751 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-config\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.394493 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-oauth-config\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.394532 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nthfk\" (UniqueName: \"kubernetes.io/projected/8b280c9f-d84d-4cda-a63a-800f116d0cca-kube-api-access-nthfk\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.394555 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-service-ca\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.394581 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-config\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.394604 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-trusted-ca-bundle\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.394622 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-serving-cert\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.394646 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-oauth-serving-cert\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.395426 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-service-ca\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.395514 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-config\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.395602 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-oauth-serving-cert\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.395886 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-trusted-ca-bundle\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.398996 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-oauth-config\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.399007 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-serving-cert\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.407728 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nthfk\" (UniqueName: \"kubernetes.io/projected/8b280c9f-d84d-4cda-a63a-800f116d0cca-kube-api-access-nthfk\") pod \"console-785bd4864d-4f7hk\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.477105 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:15:50 crc kubenswrapper[4667]: I0929 17:15:50.807387 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-785bd4864d-4f7hk"] Sep 29 17:15:51 crc kubenswrapper[4667]: I0929 17:15:51.554935 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-785bd4864d-4f7hk" event={"ID":"8b280c9f-d84d-4cda-a63a-800f116d0cca","Type":"ContainerStarted","Data":"e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538"} Sep 29 17:15:51 crc kubenswrapper[4667]: I0929 17:15:51.555102 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-785bd4864d-4f7hk" event={"ID":"8b280c9f-d84d-4cda-a63a-800f116d0cca","Type":"ContainerStarted","Data":"6e66a8c18dda8e509f1b6839f16203cb0c97886f4e4bd923f842ea17b2f8446c"} Sep 29 17:15:51 crc kubenswrapper[4667]: I0929 17:15:51.566498 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-785bd4864d-4f7hk" podStartSLOduration=1.566484305 podStartE2EDuration="1.566484305s" podCreationTimestamp="2025-09-29 17:15:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:15:51.565141802 +0000 UTC m=+380.062988572" watchObservedRunningTime="2025-09-29 17:15:51.566484305 +0000 UTC m=+380.064331074" Sep 29 17:15:56 crc kubenswrapper[4667]: I0929 17:15:56.897872 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:15:56 crc kubenswrapper[4667]: I0929 17:15:56.898246 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:16:00 crc kubenswrapper[4667]: I0929 17:16:00.477467 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:16:00 crc kubenswrapper[4667]: I0929 17:16:00.478147 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:16:00 crc kubenswrapper[4667]: I0929 17:16:00.480714 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:16:00 crc kubenswrapper[4667]: I0929 17:16:00.594267 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:16:00 crc kubenswrapper[4667]: I0929 17:16:00.625982 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-86d68f79b5-wkvqz"] Sep 29 17:16:25 crc kubenswrapper[4667]: I0929 17:16:25.652566 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-86d68f79b5-wkvqz" podUID="71707e4c-df9e-463c-a704-d1973086442e" containerName="console" containerID="cri-o://07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df" gracePeriod=15 Sep 29 17:16:25 crc kubenswrapper[4667]: I0929 17:16:25.910115 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-86d68f79b5-wkvqz_71707e4c-df9e-463c-a704-d1973086442e/console/0.log" Sep 29 17:16:25 crc kubenswrapper[4667]: I0929 17:16:25.910305 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.051882 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-serving-cert\") pod \"71707e4c-df9e-463c-a704-d1973086442e\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.051931 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-console-config\") pod \"71707e4c-df9e-463c-a704-d1973086442e\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.051967 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-trusted-ca-bundle\") pod \"71707e4c-df9e-463c-a704-d1973086442e\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.052011 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-service-ca\") pod \"71707e4c-df9e-463c-a704-d1973086442e\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.052030 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-oauth-serving-cert\") pod \"71707e4c-df9e-463c-a704-d1973086442e\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.052055 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-oauth-config\") pod \"71707e4c-df9e-463c-a704-d1973086442e\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.052108 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kps9r\" (UniqueName: \"kubernetes.io/projected/71707e4c-df9e-463c-a704-d1973086442e-kube-api-access-kps9r\") pod \"71707e4c-df9e-463c-a704-d1973086442e\" (UID: \"71707e4c-df9e-463c-a704-d1973086442e\") " Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.052623 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-service-ca" (OuterVolumeSpecName: "service-ca") pod "71707e4c-df9e-463c-a704-d1973086442e" (UID: "71707e4c-df9e-463c-a704-d1973086442e"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.052665 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "71707e4c-df9e-463c-a704-d1973086442e" (UID: "71707e4c-df9e-463c-a704-d1973086442e"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.052673 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "71707e4c-df9e-463c-a704-d1973086442e" (UID: "71707e4c-df9e-463c-a704-d1973086442e"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.052635 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-console-config" (OuterVolumeSpecName: "console-config") pod "71707e4c-df9e-463c-a704-d1973086442e" (UID: "71707e4c-df9e-463c-a704-d1973086442e"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.056330 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "71707e4c-df9e-463c-a704-d1973086442e" (UID: "71707e4c-df9e-463c-a704-d1973086442e"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.056352 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "71707e4c-df9e-463c-a704-d1973086442e" (UID: "71707e4c-df9e-463c-a704-d1973086442e"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.056433 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71707e4c-df9e-463c-a704-d1973086442e-kube-api-access-kps9r" (OuterVolumeSpecName: "kube-api-access-kps9r") pod "71707e4c-df9e-463c-a704-d1973086442e" (UID: "71707e4c-df9e-463c-a704-d1973086442e"). InnerVolumeSpecName "kube-api-access-kps9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.153569 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.153596 4667 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.153605 4667 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.153613 4667 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.153621 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kps9r\" (UniqueName: \"kubernetes.io/projected/71707e4c-df9e-463c-a704-d1973086442e-kube-api-access-kps9r\") on node \"crc\" DevicePath \"\"" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.153631 4667 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/71707e4c-df9e-463c-a704-d1973086442e-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.153639 4667 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/71707e4c-df9e-463c-a704-d1973086442e-console-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.692301 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-86d68f79b5-wkvqz_71707e4c-df9e-463c-a704-d1973086442e/console/0.log" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.692340 4667 generic.go:334] "Generic (PLEG): container finished" podID="71707e4c-df9e-463c-a704-d1973086442e" containerID="07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df" exitCode=2 Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.692365 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86d68f79b5-wkvqz" event={"ID":"71707e4c-df9e-463c-a704-d1973086442e","Type":"ContainerDied","Data":"07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df"} Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.692388 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86d68f79b5-wkvqz" event={"ID":"71707e4c-df9e-463c-a704-d1973086442e","Type":"ContainerDied","Data":"f0f88cf118de0c746ba0c380ab951e3ce3866e269d5735e0e9cd2ee5cd9cad49"} Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.692404 4667 scope.go:117] "RemoveContainer" containerID="07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.692489 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86d68f79b5-wkvqz" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.704333 4667 scope.go:117] "RemoveContainer" containerID="07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df" Sep 29 17:16:26 crc kubenswrapper[4667]: E0929 17:16:26.704618 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df\": container with ID starting with 07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df not found: ID does not exist" containerID="07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.704668 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df"} err="failed to get container status \"07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df\": rpc error: code = NotFound desc = could not find container \"07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df\": container with ID starting with 07d21d4e398a6dc8611f4f7710910ae2b40c134b6c104c2fc40ac7bb017859df not found: ID does not exist" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.711309 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-86d68f79b5-wkvqz"] Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.713628 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-86d68f79b5-wkvqz"] Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.897421 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.897467 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.897500 4667 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.897860 4667 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9ba7db32c2d86b3e6b9098efebccb5be88d2dd5c0dcb0c69d78ec2f5c309f2f"} pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 17:16:26 crc kubenswrapper[4667]: I0929 17:16:26.897916 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" containerID="cri-o://a9ba7db32c2d86b3e6b9098efebccb5be88d2dd5c0dcb0c69d78ec2f5c309f2f" gracePeriod=600 Sep 29 17:16:27 crc kubenswrapper[4667]: I0929 17:16:27.698961 4667 generic.go:334] "Generic (PLEG): container finished" podID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerID="a9ba7db32c2d86b3e6b9098efebccb5be88d2dd5c0dcb0c69d78ec2f5c309f2f" exitCode=0 Sep 29 17:16:27 crc kubenswrapper[4667]: I0929 17:16:27.699032 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerDied","Data":"a9ba7db32c2d86b3e6b9098efebccb5be88d2dd5c0dcb0c69d78ec2f5c309f2f"} Sep 29 17:16:27 crc kubenswrapper[4667]: I0929 17:16:27.699352 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"2b7b6f639c7d54c34cc1faa704af3d81c5333b8dbbef3163c511e28b934a21eb"} Sep 29 17:16:27 crc kubenswrapper[4667]: I0929 17:16:27.699372 4667 scope.go:117] "RemoveContainer" containerID="2f9695f3b47e68a1a4f7ea59a3fada8da8567592b8a6b310845aa7c90bb57657" Sep 29 17:16:27 crc kubenswrapper[4667]: I0929 17:16:27.825547 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71707e4c-df9e-463c-a704-d1973086442e" path="/var/lib/kubelet/pods/71707e4c-df9e-463c-a704-d1973086442e/volumes" Sep 29 17:18:31 crc kubenswrapper[4667]: I0929 17:18:31.953604 4667 scope.go:117] "RemoveContainer" containerID="b263e364ded51d0ec1708e37a2240252abf3690d0519b50caac9fc1eb1b58bdb" Sep 29 17:18:56 crc kubenswrapper[4667]: I0929 17:18:56.897300 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:18:56 crc kubenswrapper[4667]: I0929 17:18:56.897799 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:19:26 crc kubenswrapper[4667]: I0929 17:19:26.897162 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:19:26 crc kubenswrapper[4667]: I0929 17:19:26.898015 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:19:56 crc kubenswrapper[4667]: I0929 17:19:56.897565 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:19:56 crc kubenswrapper[4667]: I0929 17:19:56.897979 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:19:56 crc kubenswrapper[4667]: I0929 17:19:56.898018 4667 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:19:57 crc kubenswrapper[4667]: I0929 17:19:57.544120 4667 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b7b6f639c7d54c34cc1faa704af3d81c5333b8dbbef3163c511e28b934a21eb"} pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 17:19:57 crc kubenswrapper[4667]: I0929 17:19:57.544179 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" containerID="cri-o://2b7b6f639c7d54c34cc1faa704af3d81c5333b8dbbef3163c511e28b934a21eb" gracePeriod=600 Sep 29 17:19:58 crc kubenswrapper[4667]: I0929 17:19:58.549638 4667 generic.go:334] "Generic (PLEG): container finished" podID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerID="2b7b6f639c7d54c34cc1faa704af3d81c5333b8dbbef3163c511e28b934a21eb" exitCode=0 Sep 29 17:19:58 crc kubenswrapper[4667]: I0929 17:19:58.549713 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerDied","Data":"2b7b6f639c7d54c34cc1faa704af3d81c5333b8dbbef3163c511e28b934a21eb"} Sep 29 17:19:58 crc kubenswrapper[4667]: I0929 17:19:58.549972 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"24409b364de3d67a164134cbdd70a1ed09fbd088e081c52528c8d39cd38525a9"} Sep 29 17:19:58 crc kubenswrapper[4667]: I0929 17:19:58.549992 4667 scope.go:117] "RemoveContainer" containerID="a9ba7db32c2d86b3e6b9098efebccb5be88d2dd5c0dcb0c69d78ec2f5c309f2f" Sep 29 17:20:08 crc kubenswrapper[4667]: I0929 17:20:08.925872 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v"] Sep 29 17:20:08 crc kubenswrapper[4667]: E0929 17:20:08.926621 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71707e4c-df9e-463c-a704-d1973086442e" containerName="console" Sep 29 17:20:08 crc kubenswrapper[4667]: I0929 17:20:08.926636 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="71707e4c-df9e-463c-a704-d1973086442e" containerName="console" Sep 29 17:20:08 crc kubenswrapper[4667]: I0929 17:20:08.926803 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="71707e4c-df9e-463c-a704-d1973086442e" containerName="console" Sep 29 17:20:08 crc kubenswrapper[4667]: I0929 17:20:08.927935 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:08 crc kubenswrapper[4667]: I0929 17:20:08.931240 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 17:20:08 crc kubenswrapper[4667]: I0929 17:20:08.936532 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v"] Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.076756 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.076814 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.076905 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vccdv\" (UniqueName: \"kubernetes.io/projected/46414ce5-c130-445a-8c2e-a3f1e4781b9d-kube-api-access-vccdv\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.177590 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.177630 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.177692 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vccdv\" (UniqueName: \"kubernetes.io/projected/46414ce5-c130-445a-8c2e-a3f1e4781b9d-kube-api-access-vccdv\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.178079 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.178124 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.192579 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vccdv\" (UniqueName: \"kubernetes.io/projected/46414ce5-c130-445a-8c2e-a3f1e4781b9d-kube-api-access-vccdv\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.240675 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.583909 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v"] Sep 29 17:20:09 crc kubenswrapper[4667]: I0929 17:20:09.595070 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" event={"ID":"46414ce5-c130-445a-8c2e-a3f1e4781b9d","Type":"ContainerStarted","Data":"836bbd8798d68c1f210b9b85c98a75b1b6d5de74460137a79af4bb3a883f1bfd"} Sep 29 17:20:10 crc kubenswrapper[4667]: I0929 17:20:10.600988 4667 generic.go:334] "Generic (PLEG): container finished" podID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerID="4b976dbb49e4ee646f975d268b70e63d92829c402279b49f3c4559d8e5938ca3" exitCode=0 Sep 29 17:20:10 crc kubenswrapper[4667]: I0929 17:20:10.601022 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" event={"ID":"46414ce5-c130-445a-8c2e-a3f1e4781b9d","Type":"ContainerDied","Data":"4b976dbb49e4ee646f975d268b70e63d92829c402279b49f3c4559d8e5938ca3"} Sep 29 17:20:10 crc kubenswrapper[4667]: I0929 17:20:10.602210 4667 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 17:20:11 crc kubenswrapper[4667]: I0929 17:20:11.606734 4667 generic.go:334] "Generic (PLEG): container finished" podID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerID="5dc76c41d0836e081d4c6b73a0d5cd82ff2261b8a4cf10d5f73056e309d978f8" exitCode=0 Sep 29 17:20:11 crc kubenswrapper[4667]: I0929 17:20:11.606809 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" event={"ID":"46414ce5-c130-445a-8c2e-a3f1e4781b9d","Type":"ContainerDied","Data":"5dc76c41d0836e081d4c6b73a0d5cd82ff2261b8a4cf10d5f73056e309d978f8"} Sep 29 17:20:12 crc kubenswrapper[4667]: I0929 17:20:12.612881 4667 generic.go:334] "Generic (PLEG): container finished" podID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerID="cb816d42824d27387964a071f753fa3690b21a2dd7c6dcd649caf86fa11fb159" exitCode=0 Sep 29 17:20:12 crc kubenswrapper[4667]: I0929 17:20:12.612934 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" event={"ID":"46414ce5-c130-445a-8c2e-a3f1e4781b9d","Type":"ContainerDied","Data":"cb816d42824d27387964a071f753fa3690b21a2dd7c6dcd649caf86fa11fb159"} Sep 29 17:20:13 crc kubenswrapper[4667]: I0929 17:20:13.790384 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:13 crc kubenswrapper[4667]: I0929 17:20:13.931095 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-bundle\") pod \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " Sep 29 17:20:13 crc kubenswrapper[4667]: I0929 17:20:13.931238 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-util\") pod \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " Sep 29 17:20:13 crc kubenswrapper[4667]: I0929 17:20:13.931344 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vccdv\" (UniqueName: \"kubernetes.io/projected/46414ce5-c130-445a-8c2e-a3f1e4781b9d-kube-api-access-vccdv\") pod \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\" (UID: \"46414ce5-c130-445a-8c2e-a3f1e4781b9d\") " Sep 29 17:20:13 crc kubenswrapper[4667]: I0929 17:20:13.933127 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-bundle" (OuterVolumeSpecName: "bundle") pod "46414ce5-c130-445a-8c2e-a3f1e4781b9d" (UID: "46414ce5-c130-445a-8c2e-a3f1e4781b9d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:20:13 crc kubenswrapper[4667]: I0929 17:20:13.937161 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46414ce5-c130-445a-8c2e-a3f1e4781b9d-kube-api-access-vccdv" (OuterVolumeSpecName: "kube-api-access-vccdv") pod "46414ce5-c130-445a-8c2e-a3f1e4781b9d" (UID: "46414ce5-c130-445a-8c2e-a3f1e4781b9d"). InnerVolumeSpecName "kube-api-access-vccdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:20:13 crc kubenswrapper[4667]: I0929 17:20:13.941787 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-util" (OuterVolumeSpecName: "util") pod "46414ce5-c130-445a-8c2e-a3f1e4781b9d" (UID: "46414ce5-c130-445a-8c2e-a3f1e4781b9d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:20:14 crc kubenswrapper[4667]: I0929 17:20:14.032494 4667 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-util\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:14 crc kubenswrapper[4667]: I0929 17:20:14.032520 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vccdv\" (UniqueName: \"kubernetes.io/projected/46414ce5-c130-445a-8c2e-a3f1e4781b9d-kube-api-access-vccdv\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:14 crc kubenswrapper[4667]: I0929 17:20:14.032531 4667 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/46414ce5-c130-445a-8c2e-a3f1e4781b9d-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:14 crc kubenswrapper[4667]: I0929 17:20:14.625252 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" event={"ID":"46414ce5-c130-445a-8c2e-a3f1e4781b9d","Type":"ContainerDied","Data":"836bbd8798d68c1f210b9b85c98a75b1b6d5de74460137a79af4bb3a883f1bfd"} Sep 29 17:20:14 crc kubenswrapper[4667]: I0929 17:20:14.625290 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="836bbd8798d68c1f210b9b85c98a75b1b6d5de74460137a79af4bb3a883f1bfd" Sep 29 17:20:14 crc kubenswrapper[4667]: I0929 17:20:14.625348 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.788349 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4"] Sep 29 17:20:20 crc kubenswrapper[4667]: E0929 17:20:20.788678 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerName="pull" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.788689 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerName="pull" Sep 29 17:20:20 crc kubenswrapper[4667]: E0929 17:20:20.788775 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerName="extract" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.788780 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerName="extract" Sep 29 17:20:20 crc kubenswrapper[4667]: E0929 17:20:20.788790 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerName="util" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.788795 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerName="util" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.788895 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="46414ce5-c130-445a-8c2e-a3f1e4781b9d" containerName="extract" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.789206 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.791446 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-vghfs" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.793202 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.793551 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.808399 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xg6q\" (UniqueName: \"kubernetes.io/projected/04f22be1-6686-49ab-8abd-4d25428e1a8d-kube-api-access-5xg6q\") pod \"obo-prometheus-operator-7c8cf85677-x5wg4\" (UID: \"04f22be1-6686-49ab-8abd-4d25428e1a8d\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.822019 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q"] Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.822517 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.825276 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.825463 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-njczf" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.830114 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4"] Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.830760 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.837954 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q"] Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.841099 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4"] Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.843917 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4"] Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.906048 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-4mj5j"] Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.906666 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.908067 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-6q6np" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.908140 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.909340 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3a9a6102-73a2-44c8-b7fe-34e0bc6f6197-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q\" (UID: \"3a9a6102-73a2-44c8-b7fe-34e0bc6f6197\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.909408 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e01c6c8e-211e-4afd-8125-46eec155afc3-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-4mj5j\" (UID: \"e01c6c8e-211e-4afd-8125-46eec155afc3\") " pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.909458 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/677f2569-3665-45d9-bade-1aaf037a4cfb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4\" (UID: \"677f2569-3665-45d9-bade-1aaf037a4cfb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.909546 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3a9a6102-73a2-44c8-b7fe-34e0bc6f6197-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q\" (UID: \"3a9a6102-73a2-44c8-b7fe-34e0bc6f6197\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.909653 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xg6q\" (UniqueName: \"kubernetes.io/projected/04f22be1-6686-49ab-8abd-4d25428e1a8d-kube-api-access-5xg6q\") pod \"obo-prometheus-operator-7c8cf85677-x5wg4\" (UID: \"04f22be1-6686-49ab-8abd-4d25428e1a8d\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.909763 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7rzp\" (UniqueName: \"kubernetes.io/projected/e01c6c8e-211e-4afd-8125-46eec155afc3-kube-api-access-c7rzp\") pod \"observability-operator-cc5f78dfc-4mj5j\" (UID: \"e01c6c8e-211e-4afd-8125-46eec155afc3\") " pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.909797 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/677f2569-3665-45d9-bade-1aaf037a4cfb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4\" (UID: \"677f2569-3665-45d9-bade-1aaf037a4cfb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.921663 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-4mj5j"] Sep 29 17:20:20 crc kubenswrapper[4667]: I0929 17:20:20.926219 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xg6q\" (UniqueName: \"kubernetes.io/projected/04f22be1-6686-49ab-8abd-4d25428e1a8d-kube-api-access-5xg6q\") pod \"obo-prometheus-operator-7c8cf85677-x5wg4\" (UID: \"04f22be1-6686-49ab-8abd-4d25428e1a8d\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.010240 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e01c6c8e-211e-4afd-8125-46eec155afc3-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-4mj5j\" (UID: \"e01c6c8e-211e-4afd-8125-46eec155afc3\") " pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.010694 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-rphbr"] Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.010729 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/677f2569-3665-45d9-bade-1aaf037a4cfb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4\" (UID: \"677f2569-3665-45d9-bade-1aaf037a4cfb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.010768 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3a9a6102-73a2-44c8-b7fe-34e0bc6f6197-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q\" (UID: \"3a9a6102-73a2-44c8-b7fe-34e0bc6f6197\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.010809 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7rzp\" (UniqueName: \"kubernetes.io/projected/e01c6c8e-211e-4afd-8125-46eec155afc3-kube-api-access-c7rzp\") pod \"observability-operator-cc5f78dfc-4mj5j\" (UID: \"e01c6c8e-211e-4afd-8125-46eec155afc3\") " pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.010828 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/677f2569-3665-45d9-bade-1aaf037a4cfb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4\" (UID: \"677f2569-3665-45d9-bade-1aaf037a4cfb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.010872 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3a9a6102-73a2-44c8-b7fe-34e0bc6f6197-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q\" (UID: \"3a9a6102-73a2-44c8-b7fe-34e0bc6f6197\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.011664 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.013943 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-pfgl6" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.014166 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3a9a6102-73a2-44c8-b7fe-34e0bc6f6197-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q\" (UID: \"3a9a6102-73a2-44c8-b7fe-34e0bc6f6197\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.014388 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/677f2569-3665-45d9-bade-1aaf037a4cfb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4\" (UID: \"677f2569-3665-45d9-bade-1aaf037a4cfb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.014747 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e01c6c8e-211e-4afd-8125-46eec155afc3-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-4mj5j\" (UID: \"e01c6c8e-211e-4afd-8125-46eec155afc3\") " pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.015309 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3a9a6102-73a2-44c8-b7fe-34e0bc6f6197-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q\" (UID: \"3a9a6102-73a2-44c8-b7fe-34e0bc6f6197\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.024507 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-rphbr"] Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.027466 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7rzp\" (UniqueName: \"kubernetes.io/projected/e01c6c8e-211e-4afd-8125-46eec155afc3-kube-api-access-c7rzp\") pod \"observability-operator-cc5f78dfc-4mj5j\" (UID: \"e01c6c8e-211e-4afd-8125-46eec155afc3\") " pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.027695 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/677f2569-3665-45d9-bade-1aaf037a4cfb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4\" (UID: \"677f2569-3665-45d9-bade-1aaf037a4cfb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.101977 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.111608 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7ljb\" (UniqueName: \"kubernetes.io/projected/f926aaae-526c-4f52-99b3-4b799961c9d2-kube-api-access-t7ljb\") pod \"perses-operator-54bc95c9fb-rphbr\" (UID: \"f926aaae-526c-4f52-99b3-4b799961c9d2\") " pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.111662 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f926aaae-526c-4f52-99b3-4b799961c9d2-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-rphbr\" (UID: \"f926aaae-526c-4f52-99b3-4b799961c9d2\") " pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.133219 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.141406 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.212687 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7ljb\" (UniqueName: \"kubernetes.io/projected/f926aaae-526c-4f52-99b3-4b799961c9d2-kube-api-access-t7ljb\") pod \"perses-operator-54bc95c9fb-rphbr\" (UID: \"f926aaae-526c-4f52-99b3-4b799961c9d2\") " pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.212953 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f926aaae-526c-4f52-99b3-4b799961c9d2-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-rphbr\" (UID: \"f926aaae-526c-4f52-99b3-4b799961c9d2\") " pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.214188 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f926aaae-526c-4f52-99b3-4b799961c9d2-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-rphbr\" (UID: \"f926aaae-526c-4f52-99b3-4b799961c9d2\") " pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.218093 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.242700 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qjsnt"] Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.242912 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7ljb\" (UniqueName: \"kubernetes.io/projected/f926aaae-526c-4f52-99b3-4b799961c9d2-kube-api-access-t7ljb\") pod \"perses-operator-54bc95c9fb-rphbr\" (UID: \"f926aaae-526c-4f52-99b3-4b799961c9d2\") " pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.243128 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovn-controller" containerID="cri-o://813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1" gracePeriod=30 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.243214 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="sbdb" containerID="cri-o://4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd" gracePeriod=30 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.243249 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="nbdb" containerID="cri-o://3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2" gracePeriod=30 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.243279 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="northd" containerID="cri-o://585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d" gracePeriod=30 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.243307 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6" gracePeriod=30 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.243333 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kube-rbac-proxy-node" containerID="cri-o://14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546" gracePeriod=30 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.243357 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovn-acl-logging" containerID="cri-o://9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03" gracePeriod=30 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.281910 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4"] Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.320615 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(ec33c72e2a514835146e34e17fae13b295b197d02b744e5f3423472ab421d1e0): error adding pod openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): failed to send CNI request: Post \"http://dummy/cni\": EOF: StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.320678 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(ec33c72e2a514835146e34e17fae13b295b197d02b744e5f3423472ab421d1e0): error adding pod openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): failed to send CNI request: Post \"http://dummy/cni\": EOF: StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.320700 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(ec33c72e2a514835146e34e17fae13b295b197d02b744e5f3423472ab421d1e0): error adding pod openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): failed to send CNI request: Post \"http://dummy/cni\": EOF: StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.320751 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators(3a9a6102-73a2-44c8-b7fe-34e0bc6f6197)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators(3a9a6102-73a2-44c8-b7fe-34e0bc6f6197)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(ec33c72e2a514835146e34e17fae13b295b197d02b744e5f3423472ab421d1e0): error adding pod openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): failed to send CNI request: Post \\\"http://dummy/cni\\\": EOF: StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" podUID="3a9a6102-73a2-44c8-b7fe-34e0bc6f6197" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.324248 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" containerID="cri-o://1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d" gracePeriod=30 Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.344030 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(ec6f45ba357fe0721921feb195af85a8ed7e9280b3bc85283eda26a89c4c2965): error adding pod openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4 to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): failed to send CNI request: Post \"http://dummy/cni\": EOF: StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.344076 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(ec6f45ba357fe0721921feb195af85a8ed7e9280b3bc85283eda26a89c4c2965): error adding pod openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4 to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): failed to send CNI request: Post \"http://dummy/cni\": EOF: StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.344100 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(ec6f45ba357fe0721921feb195af85a8ed7e9280b3bc85283eda26a89c4c2965): error adding pod openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4 to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): failed to send CNI request: Post \"http://dummy/cni\": EOF: StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.344142 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators(677f2569-3665-45d9-bade-1aaf037a4cfb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators(677f2569-3665-45d9-bade-1aaf037a4cfb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(ec6f45ba357fe0721921feb195af85a8ed7e9280b3bc85283eda26a89c4c2965): error adding pod openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): failed to send CNI request: Post \\\"http://dummy/cni\\\": EOF: StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" podUID="677f2569-3665-45d9-bade-1aaf037a4cfb" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.354913 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.369814 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to get network status for pod sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(b329195aaef2d7123f1f95cba332059b5f6cf209edeb0da90d976693750f039d): CNI network \"\" not found" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.369870 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to get network status for pod sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(b329195aaef2d7123f1f95cba332059b5f6cf209edeb0da90d976693750f039d): CNI network \"\" not found" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.369893 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to get network status for pod sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(b329195aaef2d7123f1f95cba332059b5f6cf209edeb0da90d976693750f039d): CNI network \"\" not found" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.369928 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators(04f22be1-6686-49ab-8abd-4d25428e1a8d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators(04f22be1-6686-49ab-8abd-4d25428e1a8d)\\\": rpc error: code = Unknown desc = failed to get network status for pod sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(b329195aaef2d7123f1f95cba332059b5f6cf209edeb0da90d976693750f039d): CNI network \\\"\\\" not found\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" podUID="04f22be1-6686-49ab-8abd-4d25428e1a8d" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.387905 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(76b5d88558a1460736329c4b1431cd090632106745fa1230ceee4b66e514ebb4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.387939 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(76b5d88558a1460736329c4b1431cd090632106745fa1230ceee4b66e514ebb4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.387958 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(76b5d88558a1460736329c4b1431cd090632106745fa1230ceee4b66e514ebb4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.388001 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-54bc95c9fb-rphbr_openshift-operators(f926aaae-526c-4f52-99b3-4b799961c9d2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-54bc95c9fb-rphbr_openshift-operators(f926aaae-526c-4f52-99b3-4b799961c9d2)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(76b5d88558a1460736329c4b1431cd090632106745fa1230ceee4b66e514ebb4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" podUID="f926aaae-526c-4f52-99b3-4b799961c9d2" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.656880 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovnkube-controller/3.log" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.658505 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovn-acl-logging/0.log" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.658860 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovn-controller/0.log" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659120 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d" exitCode=0 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659142 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd" exitCode=0 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659149 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2" exitCode=0 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659156 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d" exitCode=0 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659163 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03" exitCode=143 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659169 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1" exitCode=143 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659204 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d"} Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659227 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd"} Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659236 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2"} Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659245 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d"} Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659253 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03"} Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659261 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1"} Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.659274 4667 scope.go:117] "RemoveContainer" containerID="d2a94afb0545eda590c2212482273ebde500967a5f75e3f9c068f9fa34b57415" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.664825 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/2.log" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.665196 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/1.log" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.665223 4667 generic.go:334] "Generic (PLEG): container finished" podID="788f770a-3181-4b66-981c-90ffb7fc49c0" containerID="f8a3403dd025716fcd1c6fdc3bd819d34cd5e3d986a182788c97a3df512c1dd8" exitCode=2 Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.665270 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.665526 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.665691 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2bf6m" event={"ID":"788f770a-3181-4b66-981c-90ffb7fc49c0","Type":"ContainerDied","Data":"f8a3403dd025716fcd1c6fdc3bd819d34cd5e3d986a182788c97a3df512c1dd8"} Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.665941 4667 scope.go:117] "RemoveContainer" containerID="f8a3403dd025716fcd1c6fdc3bd819d34cd5e3d986a182788c97a3df512c1dd8" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.666098 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2bf6m_openshift-multus(788f770a-3181-4b66-981c-90ffb7fc49c0)\"" pod="openshift-multus/multus-2bf6m" podUID="788f770a-3181-4b66-981c-90ffb7fc49c0" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.666359 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.666574 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.666742 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.666946 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.667095 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.667258 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: I0929 17:20:21.686804 4667 scope.go:117] "RemoveContainer" containerID="150f98cd83f5ee4326bfa0ef182ca6e470549b77c153ab419e75abce64fd6033" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.712465 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(508b502c1381d8e93f42418be363e9ba100d0eef66eb498cf6d8ec983ea337ec): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.712507 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(508b502c1381d8e93f42418be363e9ba100d0eef66eb498cf6d8ec983ea337ec): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.712526 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(508b502c1381d8e93f42418be363e9ba100d0eef66eb498cf6d8ec983ea337ec): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.712565 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-54bc95c9fb-rphbr_openshift-operators(f926aaae-526c-4f52-99b3-4b799961c9d2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-54bc95c9fb-rphbr_openshift-operators(f926aaae-526c-4f52-99b3-4b799961c9d2)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(508b502c1381d8e93f42418be363e9ba100d0eef66eb498cf6d8ec983ea337ec): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" podUID="f926aaae-526c-4f52-99b3-4b799961c9d2" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.733879 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(b2deb3763ceace1fb1a2c65f9d69da32244c21c44d4120a3b80e55c291d47ed0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.733927 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(b2deb3763ceace1fb1a2c65f9d69da32244c21c44d4120a3b80e55c291d47ed0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.733945 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(b2deb3763ceace1fb1a2c65f9d69da32244c21c44d4120a3b80e55c291d47ed0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.733981 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators(04f22be1-6686-49ab-8abd-4d25428e1a8d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators(04f22be1-6686-49ab-8abd-4d25428e1a8d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(b2deb3763ceace1fb1a2c65f9d69da32244c21c44d4120a3b80e55c291d47ed0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" podUID="04f22be1-6686-49ab-8abd-4d25428e1a8d" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.739978 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(c8a417046576e71960aef3e014b40fbed0228934b22a5d741b28d16290f033f3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.740027 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(c8a417046576e71960aef3e014b40fbed0228934b22a5d741b28d16290f033f3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.740045 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(c8a417046576e71960aef3e014b40fbed0228934b22a5d741b28d16290f033f3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.740081 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators(3a9a6102-73a2-44c8-b7fe-34e0bc6f6197)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators(3a9a6102-73a2-44c8-b7fe-34e0bc6f6197)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(c8a417046576e71960aef3e014b40fbed0228934b22a5d741b28d16290f033f3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" podUID="3a9a6102-73a2-44c8-b7fe-34e0bc6f6197" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.743664 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(326eade23e270a8013d074dc5e8551f88b51fb256099073febe119d50085d433): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.743703 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(326eade23e270a8013d074dc5e8551f88b51fb256099073febe119d50085d433): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.743730 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(326eade23e270a8013d074dc5e8551f88b51fb256099073febe119d50085d433): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.743769 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators(677f2569-3665-45d9-bade-1aaf037a4cfb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators(677f2569-3665-45d9-bade-1aaf037a4cfb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(326eade23e270a8013d074dc5e8551f88b51fb256099073febe119d50085d433): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" podUID="677f2569-3665-45d9-bade-1aaf037a4cfb" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.927899 4667 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd is running failed: container process not found" containerID="4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.927899 4667 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2 is running failed: container process not found" containerID="3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.928218 4667 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd is running failed: container process not found" containerID="4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.928330 4667 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2 is running failed: container process not found" containerID="3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.928581 4667 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd is running failed: container process not found" containerID="4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.928606 4667 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="sbdb" Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.928652 4667 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2 is running failed: container process not found" containerID="3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Sep 29 17:20:21 crc kubenswrapper[4667]: E0929 17:20:21.928666 4667 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="nbdb" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.476108 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovn-acl-logging/0.log" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.476664 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovn-controller/0.log" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.477004 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.510965 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wqqpn"] Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511159 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511175 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511183 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovn-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511189 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovn-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511195 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="northd" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511200 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="northd" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511210 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kube-rbac-proxy-ovn-metrics" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511216 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kube-rbac-proxy-ovn-metrics" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511223 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kubecfg-setup" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511228 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kubecfg-setup" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511239 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="sbdb" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511244 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="sbdb" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511251 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511256 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511263 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511268 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511274 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kube-rbac-proxy-node" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511280 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kube-rbac-proxy-node" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511286 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="nbdb" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511291 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="nbdb" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511299 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovn-acl-logging" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511304 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovn-acl-logging" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511313 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511318 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511397 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511405 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="nbdb" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511413 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511420 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511427 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovn-acl-logging" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511435 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="northd" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511440 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovn-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511446 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="sbdb" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511453 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511459 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kube-rbac-proxy-node" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511465 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="kube-rbac-proxy-ovn-metrics" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.511553 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511560 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.511651 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerName="ovnkube-controller" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.513071 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629168 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-systemd\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629224 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-ovn\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629247 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-netns\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629282 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-config\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629298 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-openvswitch\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629313 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovn-node-metrics-cert\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629327 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-etc-openvswitch\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629340 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-kubelet\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629365 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-ovn-kubernetes\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629379 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-var-lib-openvswitch\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629394 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629414 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-systemd-units\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629430 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-netd\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629444 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-env-overrides\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629446 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629457 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-bin\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629474 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqs9s\" (UniqueName: \"kubernetes.io/projected/6250a9ff-80f5-44d8-90f6-40e77106af6c-kube-api-access-bqs9s\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629494 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-node-log\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629508 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-log-socket\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629522 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-slash\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629542 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-script-lib\") pod \"6250a9ff-80f5-44d8-90f6-40e77106af6c\" (UID: \"6250a9ff-80f5-44d8-90f6-40e77106af6c\") " Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629628 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6032dc92-2824-464c-9c16-71ee6f4b46db-ovnkube-config\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629646 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-cni-netd\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629663 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-cni-bin\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629681 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-log-socket\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629697 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-kubelet\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629711 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-run-openvswitch\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629741 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-run-netns\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629754 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-run-ovn\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629769 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6032dc92-2824-464c-9c16-71ee6f4b46db-ovn-node-metrics-cert\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629783 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtw4w\" (UniqueName: \"kubernetes.io/projected/6032dc92-2824-464c-9c16-71ee6f4b46db-kube-api-access-mtw4w\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629797 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-node-log\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629834 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-var-lib-openvswitch\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629867 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-run-systemd\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629882 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-systemd-units\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629896 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-run-ovn-kubernetes\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629909 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-slash\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629897 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-log-socket" (OuterVolumeSpecName: "log-socket") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629923 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-etc-openvswitch\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629938 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629953 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6032dc92-2824-464c-9c16-71ee6f4b46db-ovnkube-script-lib\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.629978 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6032dc92-2824-464c-9c16-71ee6f4b46db-env-overrides\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630012 4667 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630021 4667 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-log-socket\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630060 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-slash" (OuterVolumeSpecName: "host-slash") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630323 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630428 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630453 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630468 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630483 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630500 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630515 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630530 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630545 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630694 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.630761 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.631038 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.631883 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.631965 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-node-log" (OuterVolumeSpecName: "node-log") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.635735 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.635927 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6250a9ff-80f5-44d8-90f6-40e77106af6c-kube-api-access-bqs9s" (OuterVolumeSpecName: "kube-api-access-bqs9s") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "kube-api-access-bqs9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.642355 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "6250a9ff-80f5-44d8-90f6-40e77106af6c" (UID: "6250a9ff-80f5-44d8-90f6-40e77106af6c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.671785 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovn-acl-logging/0.log" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.672199 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qjsnt_6250a9ff-80f5-44d8-90f6-40e77106af6c/ovn-controller/0.log" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.672442 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6" exitCode=0 Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.672467 4667 generic.go:334] "Generic (PLEG): container finished" podID="6250a9ff-80f5-44d8-90f6-40e77106af6c" containerID="14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546" exitCode=0 Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.672503 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.672583 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6"} Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.672682 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546"} Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.672752 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qjsnt" event={"ID":"6250a9ff-80f5-44d8-90f6-40e77106af6c","Type":"ContainerDied","Data":"87837f7510257125171781c62deb4919c2d91e84936b76106ade551402416775"} Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.672760 4667 scope.go:117] "RemoveContainer" containerID="1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.674056 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/2.log" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.690454 4667 scope.go:117] "RemoveContainer" containerID="4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.697880 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qjsnt"] Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.700416 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qjsnt"] Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.705361 4667 scope.go:117] "RemoveContainer" containerID="3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.717136 4667 scope.go:117] "RemoveContainer" containerID="585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.727574 4667 scope.go:117] "RemoveContainer" containerID="9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730777 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-node-log\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730835 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-var-lib-openvswitch\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730890 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-run-systemd\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730907 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-systemd-units\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730912 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-node-log\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730922 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-run-ovn-kubernetes\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730952 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-run-ovn-kubernetes\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730971 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-slash\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730983 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-var-lib-openvswitch\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.730999 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-etc-openvswitch\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731004 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-run-systemd\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731025 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-systemd-units\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731024 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731043 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-slash\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731054 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6032dc92-2824-464c-9c16-71ee6f4b46db-ovnkube-script-lib\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731067 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-etc-openvswitch\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731087 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731134 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6032dc92-2824-464c-9c16-71ee6f4b46db-env-overrides\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731168 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-cni-netd\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731181 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6032dc92-2824-464c-9c16-71ee6f4b46db-ovnkube-config\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731214 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-cni-bin\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731240 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-log-socket\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731256 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-kubelet\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731270 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-run-openvswitch\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731277 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-cni-netd\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731304 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-run-netns\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731324 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-run-ovn\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731343 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-run-openvswitch\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731375 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-run-netns\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731379 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-log-socket\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731398 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-run-ovn\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731324 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-kubelet\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731343 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6032dc92-2824-464c-9c16-71ee6f4b46db-ovn-node-metrics-cert\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731436 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6032dc92-2824-464c-9c16-71ee6f4b46db-host-cni-bin\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731441 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtw4w\" (UniqueName: \"kubernetes.io/projected/6032dc92-2824-464c-9c16-71ee6f4b46db-kube-api-access-mtw4w\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731535 4667 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-systemd-units\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731545 4667 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731554 4667 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-env-overrides\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731563 4667 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731575 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqs9s\" (UniqueName: \"kubernetes.io/projected/6250a9ff-80f5-44d8-90f6-40e77106af6c-kube-api-access-bqs9s\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731584 4667 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-node-log\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731593 4667 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-slash\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731600 4667 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731609 4667 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-systemd\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731617 4667 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-run-ovn\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731624 4667 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-netns\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731632 4667 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731641 4667 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6250a9ff-80f5-44d8-90f6-40e77106af6c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731648 4667 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731656 4667 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-kubelet\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731663 4667 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731671 4667 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731679 4667 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6250a9ff-80f5-44d8-90f6-40e77106af6c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.731661 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6032dc92-2824-464c-9c16-71ee6f4b46db-env-overrides\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.732261 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6032dc92-2824-464c-9c16-71ee6f4b46db-ovnkube-config\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.732403 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6032dc92-2824-464c-9c16-71ee6f4b46db-ovnkube-script-lib\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.734090 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6032dc92-2824-464c-9c16-71ee6f4b46db-ovn-node-metrics-cert\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.744677 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtw4w\" (UniqueName: \"kubernetes.io/projected/6032dc92-2824-464c-9c16-71ee6f4b46db-kube-api-access-mtw4w\") pod \"ovnkube-node-wqqpn\" (UID: \"6032dc92-2824-464c-9c16-71ee6f4b46db\") " pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.747647 4667 scope.go:117] "RemoveContainer" containerID="14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.757055 4667 scope.go:117] "RemoveContainer" containerID="9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.766703 4667 scope.go:117] "RemoveContainer" containerID="813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.777249 4667 scope.go:117] "RemoveContainer" containerID="dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.788429 4667 scope.go:117] "RemoveContainer" containerID="1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.788707 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d\": container with ID starting with 1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d not found: ID does not exist" containerID="1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.788750 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d"} err="failed to get container status \"1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d\": rpc error: code = NotFound desc = could not find container \"1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d\": container with ID starting with 1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.788768 4667 scope.go:117] "RemoveContainer" containerID="4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.789042 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\": container with ID starting with 4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd not found: ID does not exist" containerID="4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.789079 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd"} err="failed to get container status \"4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\": rpc error: code = NotFound desc = could not find container \"4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\": container with ID starting with 4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.789098 4667 scope.go:117] "RemoveContainer" containerID="3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.789311 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\": container with ID starting with 3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2 not found: ID does not exist" containerID="3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.789337 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2"} err="failed to get container status \"3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\": rpc error: code = NotFound desc = could not find container \"3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\": container with ID starting with 3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.789352 4667 scope.go:117] "RemoveContainer" containerID="585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.789612 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\": container with ID starting with 585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d not found: ID does not exist" containerID="585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.789631 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d"} err="failed to get container status \"585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\": rpc error: code = NotFound desc = could not find container \"585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\": container with ID starting with 585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.789643 4667 scope.go:117] "RemoveContainer" containerID="9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.789868 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\": container with ID starting with 9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6 not found: ID does not exist" containerID="9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.789892 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6"} err="failed to get container status \"9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\": rpc error: code = NotFound desc = could not find container \"9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\": container with ID starting with 9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.789906 4667 scope.go:117] "RemoveContainer" containerID="14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.790106 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\": container with ID starting with 14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546 not found: ID does not exist" containerID="14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.790128 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546"} err="failed to get container status \"14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\": rpc error: code = NotFound desc = could not find container \"14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\": container with ID starting with 14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.790141 4667 scope.go:117] "RemoveContainer" containerID="9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.790381 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\": container with ID starting with 9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03 not found: ID does not exist" containerID="9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.790412 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03"} err="failed to get container status \"9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\": rpc error: code = NotFound desc = could not find container \"9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\": container with ID starting with 9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.790431 4667 scope.go:117] "RemoveContainer" containerID="813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.790672 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\": container with ID starting with 813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1 not found: ID does not exist" containerID="813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.790692 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1"} err="failed to get container status \"813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\": rpc error: code = NotFound desc = could not find container \"813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\": container with ID starting with 813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.790706 4667 scope.go:117] "RemoveContainer" containerID="dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a" Sep 29 17:20:22 crc kubenswrapper[4667]: E0929 17:20:22.790937 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\": container with ID starting with dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a not found: ID does not exist" containerID="dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.790958 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a"} err="failed to get container status \"dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\": rpc error: code = NotFound desc = could not find container \"dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\": container with ID starting with dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.790971 4667 scope.go:117] "RemoveContainer" containerID="1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.791181 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d"} err="failed to get container status \"1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d\": rpc error: code = NotFound desc = could not find container \"1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d\": container with ID starting with 1cd6e3085d5073992a37bd7da3d0b72ae3257324b6436b1c9a9693f4bd2acc0d not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.791201 4667 scope.go:117] "RemoveContainer" containerID="4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.791403 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd"} err="failed to get container status \"4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\": rpc error: code = NotFound desc = could not find container \"4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd\": container with ID starting with 4a21af4d871575825f04a144b3bb13c6e554dd444ebc670ca62019fa716ef6cd not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.791420 4667 scope.go:117] "RemoveContainer" containerID="3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.791595 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2"} err="failed to get container status \"3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\": rpc error: code = NotFound desc = could not find container \"3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2\": container with ID starting with 3e985e783b97aed7f7034f4fd06652558292d079206d1a0e40c145c2915cb0d2 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.791616 4667 scope.go:117] "RemoveContainer" containerID="585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.791787 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d"} err="failed to get container status \"585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\": rpc error: code = NotFound desc = could not find container \"585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d\": container with ID starting with 585f6f4b1c35bc16402ec20dc1428c78246e23bcb72868f6628226aca4fd569d not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.791813 4667 scope.go:117] "RemoveContainer" containerID="9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.792051 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6"} err="failed to get container status \"9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\": rpc error: code = NotFound desc = could not find container \"9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6\": container with ID starting with 9b25e7445faa2c254b74ca214b1ae579e92e7275c464172ead0eecf1dd791df6 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.792068 4667 scope.go:117] "RemoveContainer" containerID="14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.792307 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546"} err="failed to get container status \"14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\": rpc error: code = NotFound desc = could not find container \"14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546\": container with ID starting with 14056776d3574367546941b96f4054cae78d70151c4c2f6b95c1875339be4546 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.792327 4667 scope.go:117] "RemoveContainer" containerID="9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.792547 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03"} err="failed to get container status \"9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\": rpc error: code = NotFound desc = could not find container \"9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03\": container with ID starting with 9e4dff492cad269b1936815ec20cf34719b7c032758ab0bde1d43753b1f24f03 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.792565 4667 scope.go:117] "RemoveContainer" containerID="813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.792790 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1"} err="failed to get container status \"813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\": rpc error: code = NotFound desc = could not find container \"813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1\": container with ID starting with 813c36cdf98b57144bd72ca19d0a792a1235ea25c263fb14e20816f9d2abfcb1 not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.792811 4667 scope.go:117] "RemoveContainer" containerID="dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.793051 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a"} err="failed to get container status \"dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\": rpc error: code = NotFound desc = could not find container \"dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a\": container with ID starting with dc2646b0d8eff9d993b3254f5e5889a2be43a89f2c94c496dff359c0b988512a not found: ID does not exist" Sep 29 17:20:22 crc kubenswrapper[4667]: I0929 17:20:22.822596 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:23 crc kubenswrapper[4667]: I0929 17:20:23.680492 4667 generic.go:334] "Generic (PLEG): container finished" podID="6032dc92-2824-464c-9c16-71ee6f4b46db" containerID="1df983ce21b2cfa11c96d33aa9fffc84d7aed5ae1100e84fb7d7c2c4a8f5dcc9" exitCode=0 Sep 29 17:20:23 crc kubenswrapper[4667]: I0929 17:20:23.680518 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerDied","Data":"1df983ce21b2cfa11c96d33aa9fffc84d7aed5ae1100e84fb7d7c2c4a8f5dcc9"} Sep 29 17:20:23 crc kubenswrapper[4667]: I0929 17:20:23.680781 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerStarted","Data":"2afb7b31d705a460f6e6738e34662e5fb274ac7d8bfbc2c1302c9c98b168e437"} Sep 29 17:20:23 crc kubenswrapper[4667]: I0929 17:20:23.821993 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6250a9ff-80f5-44d8-90f6-40e77106af6c" path="/var/lib/kubelet/pods/6250a9ff-80f5-44d8-90f6-40e77106af6c/volumes" Sep 29 17:20:24 crc kubenswrapper[4667]: I0929 17:20:24.688432 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerStarted","Data":"c15638f6ca2a7d42f04f406d0c95c6be93de083de5d5a9f0cab51e5e77b20de1"} Sep 29 17:20:24 crc kubenswrapper[4667]: I0929 17:20:24.688496 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerStarted","Data":"5863223807d1dd303541f82bd1031216279f34eded30b89c4fddbac61d2a3e58"} Sep 29 17:20:24 crc kubenswrapper[4667]: I0929 17:20:24.688510 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerStarted","Data":"7b2e5dcb4b8268b610a470e5c7076f4787eb59fcfe48248e18d234cfda594ca1"} Sep 29 17:20:24 crc kubenswrapper[4667]: I0929 17:20:24.688518 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerStarted","Data":"d7c4b19fbb25c2fc0f8a68649969ae4caed43b94fad1a7f08d4a0a0e52aaae25"} Sep 29 17:20:24 crc kubenswrapper[4667]: I0929 17:20:24.688527 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerStarted","Data":"61d715f27892d7f7ee431a02f61d3941b76e1bfdab8770dcff01a98306e5239a"} Sep 29 17:20:24 crc kubenswrapper[4667]: I0929 17:20:24.688534 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerStarted","Data":"fe2cd87b2711b3252a50b7a299e729c92d78f4efa1eed6e632cefe20c6ae27d0"} Sep 29 17:20:26 crc kubenswrapper[4667]: I0929 17:20:26.700450 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerStarted","Data":"5a24c59456190e17b6c0ccd608bda8644fe7c45227bdeb052e88d48db407ea8d"} Sep 29 17:20:28 crc kubenswrapper[4667]: I0929 17:20:28.712685 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" event={"ID":"6032dc92-2824-464c-9c16-71ee6f4b46db","Type":"ContainerStarted","Data":"6d3f722b4c8b6a9363a7216d177a294c1eefc7dbcf97e616648ef222a63a8e35"} Sep 29 17:20:28 crc kubenswrapper[4667]: I0929 17:20:28.713107 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:28 crc kubenswrapper[4667]: I0929 17:20:28.713119 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:28 crc kubenswrapper[4667]: I0929 17:20:28.735650 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" podStartSLOduration=6.735638707 podStartE2EDuration="6.735638707s" podCreationTimestamp="2025-09-29 17:20:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:20:28.732952402 +0000 UTC m=+657.230799172" watchObservedRunningTime="2025-09-29 17:20:28.735638707 +0000 UTC m=+657.233485477" Sep 29 17:20:28 crc kubenswrapper[4667]: I0929 17:20:28.738930 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:29 crc kubenswrapper[4667]: I0929 17:20:29.716526 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:29 crc kubenswrapper[4667]: I0929 17:20:29.738587 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:32 crc kubenswrapper[4667]: I0929 17:20:32.815650 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:32 crc kubenswrapper[4667]: I0929 17:20:32.816385 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:32 crc kubenswrapper[4667]: E0929 17:20:32.836198 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(3466f3898f14cff39187d0aeccf42444f86a28a7a49f390f1a88a2f08044528e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:32 crc kubenswrapper[4667]: E0929 17:20:32.836374 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(3466f3898f14cff39187d0aeccf42444f86a28a7a49f390f1a88a2f08044528e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:32 crc kubenswrapper[4667]: E0929 17:20:32.836393 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(3466f3898f14cff39187d0aeccf42444f86a28a7a49f390f1a88a2f08044528e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:32 crc kubenswrapper[4667]: E0929 17:20:32.836433 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators(04f22be1-6686-49ab-8abd-4d25428e1a8d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators(04f22be1-6686-49ab-8abd-4d25428e1a8d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(3466f3898f14cff39187d0aeccf42444f86a28a7a49f390f1a88a2f08044528e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" podUID="04f22be1-6686-49ab-8abd-4d25428e1a8d" Sep 29 17:20:33 crc kubenswrapper[4667]: I0929 17:20:33.815693 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:33 crc kubenswrapper[4667]: I0929 17:20:33.816111 4667 scope.go:117] "RemoveContainer" containerID="f8a3403dd025716fcd1c6fdc3bd819d34cd5e3d986a182788c97a3df512c1dd8" Sep 29 17:20:33 crc kubenswrapper[4667]: I0929 17:20:33.816185 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:33 crc kubenswrapper[4667]: E0929 17:20:33.816341 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2bf6m_openshift-multus(788f770a-3181-4b66-981c-90ffb7fc49c0)\"" pod="openshift-multus/multus-2bf6m" podUID="788f770a-3181-4b66-981c-90ffb7fc49c0" Sep 29 17:20:33 crc kubenswrapper[4667]: E0929 17:20:33.835005 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(2f35ba097ba033c71dba64a14ea91009d74c294e6ef8122a9903e7a9d81047aa): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:33 crc kubenswrapper[4667]: E0929 17:20:33.835052 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(2f35ba097ba033c71dba64a14ea91009d74c294e6ef8122a9903e7a9d81047aa): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:33 crc kubenswrapper[4667]: E0929 17:20:33.835087 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(2f35ba097ba033c71dba64a14ea91009d74c294e6ef8122a9903e7a9d81047aa): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:33 crc kubenswrapper[4667]: E0929 17:20:33.835123 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-54bc95c9fb-rphbr_openshift-operators(f926aaae-526c-4f52-99b3-4b799961c9d2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-54bc95c9fb-rphbr_openshift-operators(f926aaae-526c-4f52-99b3-4b799961c9d2)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(2f35ba097ba033c71dba64a14ea91009d74c294e6ef8122a9903e7a9d81047aa): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" podUID="f926aaae-526c-4f52-99b3-4b799961c9d2" Sep 29 17:20:34 crc kubenswrapper[4667]: I0929 17:20:34.814910 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:34 crc kubenswrapper[4667]: I0929 17:20:34.815396 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:34 crc kubenswrapper[4667]: E0929 17:20:34.839583 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(777bcfa301367289916860e259a8a486f7a319a9072b217fbbae7bdde1faf911): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:34 crc kubenswrapper[4667]: E0929 17:20:34.839635 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(777bcfa301367289916860e259a8a486f7a319a9072b217fbbae7bdde1faf911): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:34 crc kubenswrapper[4667]: E0929 17:20:34.839669 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(777bcfa301367289916860e259a8a486f7a319a9072b217fbbae7bdde1faf911): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:34 crc kubenswrapper[4667]: E0929 17:20:34.839714 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators(677f2569-3665-45d9-bade-1aaf037a4cfb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators(677f2569-3665-45d9-bade-1aaf037a4cfb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(777bcfa301367289916860e259a8a486f7a319a9072b217fbbae7bdde1faf911): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" podUID="677f2569-3665-45d9-bade-1aaf037a4cfb" Sep 29 17:20:36 crc kubenswrapper[4667]: I0929 17:20:36.815103 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:36 crc kubenswrapper[4667]: I0929 17:20:36.815682 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:36 crc kubenswrapper[4667]: E0929 17:20:36.831579 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(21aa51c2aea72bf78969ad286125da32a441b1da6cf6b21c59276afa471edc3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:36 crc kubenswrapper[4667]: E0929 17:20:36.831636 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(21aa51c2aea72bf78969ad286125da32a441b1da6cf6b21c59276afa471edc3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:36 crc kubenswrapper[4667]: E0929 17:20:36.831659 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(21aa51c2aea72bf78969ad286125da32a441b1da6cf6b21c59276afa471edc3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:36 crc kubenswrapper[4667]: E0929 17:20:36.831702 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators(3a9a6102-73a2-44c8-b7fe-34e0bc6f6197)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators(3a9a6102-73a2-44c8-b7fe-34e0bc6f6197)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_openshift-operators_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197_0(21aa51c2aea72bf78969ad286125da32a441b1da6cf6b21c59276afa471edc3d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" podUID="3a9a6102-73a2-44c8-b7fe-34e0bc6f6197" Sep 29 17:20:43 crc kubenswrapper[4667]: I0929 17:20:43.815695 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:43 crc kubenswrapper[4667]: I0929 17:20:43.816385 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:43 crc kubenswrapper[4667]: E0929 17:20:43.836691 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(a68aaa29ddccf444de8aeacae4eb6a6ec8992ae4c5daa6857958cf217e1a0b42): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:43 crc kubenswrapper[4667]: E0929 17:20:43.836766 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(a68aaa29ddccf444de8aeacae4eb6a6ec8992ae4c5daa6857958cf217e1a0b42): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:43 crc kubenswrapper[4667]: E0929 17:20:43.836786 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(a68aaa29ddccf444de8aeacae4eb6a6ec8992ae4c5daa6857958cf217e1a0b42): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:43 crc kubenswrapper[4667]: E0929 17:20:43.836826 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators(04f22be1-6686-49ab-8abd-4d25428e1a8d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators(04f22be1-6686-49ab-8abd-4d25428e1a8d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-7c8cf85677-x5wg4_openshift-operators_04f22be1-6686-49ab-8abd-4d25428e1a8d_0(a68aaa29ddccf444de8aeacae4eb6a6ec8992ae4c5daa6857958cf217e1a0b42): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" podUID="04f22be1-6686-49ab-8abd-4d25428e1a8d" Sep 29 17:20:44 crc kubenswrapper[4667]: I0929 17:20:44.815131 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:44 crc kubenswrapper[4667]: I0929 17:20:44.815737 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:44 crc kubenswrapper[4667]: E0929 17:20:44.832949 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(f7542b87fead4cb71196c2332c0fd859e1e549d387b4b3eef41e7d02845dfae3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:44 crc kubenswrapper[4667]: E0929 17:20:44.833326 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(f7542b87fead4cb71196c2332c0fd859e1e549d387b4b3eef41e7d02845dfae3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:44 crc kubenswrapper[4667]: E0929 17:20:44.833394 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(f7542b87fead4cb71196c2332c0fd859e1e549d387b4b3eef41e7d02845dfae3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:44 crc kubenswrapper[4667]: E0929 17:20:44.833479 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-54bc95c9fb-rphbr_openshift-operators(f926aaae-526c-4f52-99b3-4b799961c9d2)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-54bc95c9fb-rphbr_openshift-operators(f926aaae-526c-4f52-99b3-4b799961c9d2)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-54bc95c9fb-rphbr_openshift-operators_f926aaae-526c-4f52-99b3-4b799961c9d2_0(f7542b87fead4cb71196c2332c0fd859e1e549d387b4b3eef41e7d02845dfae3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" podUID="f926aaae-526c-4f52-99b3-4b799961c9d2" Sep 29 17:20:46 crc kubenswrapper[4667]: I0929 17:20:46.814983 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:46 crc kubenswrapper[4667]: I0929 17:20:46.815438 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:46 crc kubenswrapper[4667]: E0929 17:20:46.833467 4667 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(2304696e658f8c16eca67a5a9ec7389cd7b493f797517636dedc6738b37a4cc4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Sep 29 17:20:46 crc kubenswrapper[4667]: E0929 17:20:46.833518 4667 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(2304696e658f8c16eca67a5a9ec7389cd7b493f797517636dedc6738b37a4cc4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:46 crc kubenswrapper[4667]: E0929 17:20:46.833537 4667 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(2304696e658f8c16eca67a5a9ec7389cd7b493f797517636dedc6738b37a4cc4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:46 crc kubenswrapper[4667]: E0929 17:20:46.833577 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators(677f2569-3665-45d9-bade-1aaf037a4cfb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators(677f2569-3665-45d9-bade-1aaf037a4cfb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_openshift-operators_677f2569-3665-45d9-bade-1aaf037a4cfb_0(2304696e658f8c16eca67a5a9ec7389cd7b493f797517636dedc6738b37a4cc4): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" podUID="677f2569-3665-45d9-bade-1aaf037a4cfb" Sep 29 17:20:47 crc kubenswrapper[4667]: I0929 17:20:47.815871 4667 scope.go:117] "RemoveContainer" containerID="f8a3403dd025716fcd1c6fdc3bd819d34cd5e3d986a182788c97a3df512c1dd8" Sep 29 17:20:48 crc kubenswrapper[4667]: I0929 17:20:48.797422 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2bf6m_788f770a-3181-4b66-981c-90ffb7fc49c0/kube-multus/2.log" Sep 29 17:20:48 crc kubenswrapper[4667]: I0929 17:20:48.797660 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2bf6m" event={"ID":"788f770a-3181-4b66-981c-90ffb7fc49c0","Type":"ContainerStarted","Data":"74b5393baa1f8375f6cd4dba744173f20a0efe208354b76825fffd2a46d47822"} Sep 29 17:20:49 crc kubenswrapper[4667]: I0929 17:20:49.370269 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-4mj5j"] Sep 29 17:20:49 crc kubenswrapper[4667]: W0929 17:20:49.382423 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode01c6c8e_211e_4afd_8125_46eec155afc3.slice/crio-2b3fe6a7762face59701b1294aaaf844b6ffa21516d2fced4fac3ca7dc865767 WatchSource:0}: Error finding container 2b3fe6a7762face59701b1294aaaf844b6ffa21516d2fced4fac3ca7dc865767: Status 404 returned error can't find the container with id 2b3fe6a7762face59701b1294aaaf844b6ffa21516d2fced4fac3ca7dc865767 Sep 29 17:20:49 crc kubenswrapper[4667]: I0929 17:20:49.803202 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" event={"ID":"e01c6c8e-211e-4afd-8125-46eec155afc3","Type":"ContainerStarted","Data":"2b3fe6a7762face59701b1294aaaf844b6ffa21516d2fced4fac3ca7dc865767"} Sep 29 17:20:49 crc kubenswrapper[4667]: I0929 17:20:49.815552 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:49 crc kubenswrapper[4667]: I0929 17:20:49.816040 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" Sep 29 17:20:50 crc kubenswrapper[4667]: I0929 17:20:50.140538 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q"] Sep 29 17:20:50 crc kubenswrapper[4667]: W0929 17:20:50.144522 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a9a6102_73a2_44c8_b7fe_34e0bc6f6197.slice/crio-2ee55eb0b7f2b3c15ae55e04332e544e3f73c2d722025d1d0be9d7c21e71404a WatchSource:0}: Error finding container 2ee55eb0b7f2b3c15ae55e04332e544e3f73c2d722025d1d0be9d7c21e71404a: Status 404 returned error can't find the container with id 2ee55eb0b7f2b3c15ae55e04332e544e3f73c2d722025d1d0be9d7c21e71404a Sep 29 17:20:50 crc kubenswrapper[4667]: I0929 17:20:50.809177 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" event={"ID":"3a9a6102-73a2-44c8-b7fe-34e0bc6f6197","Type":"ContainerStarted","Data":"2ee55eb0b7f2b3c15ae55e04332e544e3f73c2d722025d1d0be9d7c21e71404a"} Sep 29 17:20:52 crc kubenswrapper[4667]: I0929 17:20:52.844116 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wqqpn" Sep 29 17:20:55 crc kubenswrapper[4667]: I0929 17:20:55.852915 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" event={"ID":"e01c6c8e-211e-4afd-8125-46eec155afc3","Type":"ContainerStarted","Data":"ab7591d35b9523618bd07b3c10a499897d767556c625015d48f7871effe23358"} Sep 29 17:20:55 crc kubenswrapper[4667]: I0929 17:20:55.853114 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:55 crc kubenswrapper[4667]: I0929 17:20:55.854936 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" Sep 29 17:20:55 crc kubenswrapper[4667]: I0929 17:20:55.855029 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" event={"ID":"3a9a6102-73a2-44c8-b7fe-34e0bc6f6197","Type":"ContainerStarted","Data":"459221bcce8fd389ad9b15877d594ae525754dfe247cdde6214fa873a52e1cbb"} Sep 29 17:20:55 crc kubenswrapper[4667]: I0929 17:20:55.865306 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-4mj5j" podStartSLOduration=29.706832857 podStartE2EDuration="35.865293498s" podCreationTimestamp="2025-09-29 17:20:20 +0000 UTC" firstStartedPulling="2025-09-29 17:20:49.384477202 +0000 UTC m=+677.882323972" lastFinishedPulling="2025-09-29 17:20:55.542937844 +0000 UTC m=+684.040784613" observedRunningTime="2025-09-29 17:20:55.86388926 +0000 UTC m=+684.361736029" watchObservedRunningTime="2025-09-29 17:20:55.865293498 +0000 UTC m=+684.363140266" Sep 29 17:20:55 crc kubenswrapper[4667]: I0929 17:20:55.878494 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q" podStartSLOduration=30.498696361 podStartE2EDuration="35.878479923s" podCreationTimestamp="2025-09-29 17:20:20 +0000 UTC" firstStartedPulling="2025-09-29 17:20:50.14588686 +0000 UTC m=+678.643733629" lastFinishedPulling="2025-09-29 17:20:55.525670423 +0000 UTC m=+684.023517191" observedRunningTime="2025-09-29 17:20:55.87558735 +0000 UTC m=+684.373434118" watchObservedRunningTime="2025-09-29 17:20:55.878479923 +0000 UTC m=+684.376326692" Sep 29 17:20:56 crc kubenswrapper[4667]: I0929 17:20:56.814873 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:56 crc kubenswrapper[4667]: I0929 17:20:56.815495 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:57 crc kubenswrapper[4667]: I0929 17:20:57.148434 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-rphbr"] Sep 29 17:20:57 crc kubenswrapper[4667]: W0929 17:20:57.152794 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf926aaae_526c_4f52_99b3_4b799961c9d2.slice/crio-1736881faa86afe49fa16a1d60ef9a8c174e69b13efe71c6510b521f75f22d31 WatchSource:0}: Error finding container 1736881faa86afe49fa16a1d60ef9a8c174e69b13efe71c6510b521f75f22d31: Status 404 returned error can't find the container with id 1736881faa86afe49fa16a1d60ef9a8c174e69b13efe71c6510b521f75f22d31 Sep 29 17:20:57 crc kubenswrapper[4667]: I0929 17:20:57.815438 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:57 crc kubenswrapper[4667]: I0929 17:20:57.816047 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" Sep 29 17:20:57 crc kubenswrapper[4667]: I0929 17:20:57.867992 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" event={"ID":"f926aaae-526c-4f52-99b3-4b799961c9d2","Type":"ContainerStarted","Data":"1736881faa86afe49fa16a1d60ef9a8c174e69b13efe71c6510b521f75f22d31"} Sep 29 17:20:58 crc kubenswrapper[4667]: I0929 17:20:58.154444 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4"] Sep 29 17:20:58 crc kubenswrapper[4667]: I0929 17:20:58.814651 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:58 crc kubenswrapper[4667]: I0929 17:20:58.815213 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" Sep 29 17:20:58 crc kubenswrapper[4667]: I0929 17:20:58.873961 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" event={"ID":"677f2569-3665-45d9-bade-1aaf037a4cfb","Type":"ContainerStarted","Data":"52bc19d14c1390df729cdf90cbd521fe2d664eb4562fed0ee08f9ecee6f5c703"} Sep 29 17:20:58 crc kubenswrapper[4667]: I0929 17:20:58.874123 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" event={"ID":"677f2569-3665-45d9-bade-1aaf037a4cfb","Type":"ContainerStarted","Data":"8eae9643e8706b980704259ca5edb8472f5d5d01148b8832f2a68fa6aef05388"} Sep 29 17:20:58 crc kubenswrapper[4667]: I0929 17:20:58.878442 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" event={"ID":"f926aaae-526c-4f52-99b3-4b799961c9d2","Type":"ContainerStarted","Data":"aa89e742e5fe377e9a66448cb01787f17d10339490633e9a96677486315ffa4d"} Sep 29 17:20:58 crc kubenswrapper[4667]: I0929 17:20:58.878563 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:20:58 crc kubenswrapper[4667]: I0929 17:20:58.887881 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-646f8974b4-r79h4" podStartSLOduration=38.887866565 podStartE2EDuration="38.887866565s" podCreationTimestamp="2025-09-29 17:20:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:20:58.887463595 +0000 UTC m=+687.385310364" watchObservedRunningTime="2025-09-29 17:20:58.887866565 +0000 UTC m=+687.385713334" Sep 29 17:20:58 crc kubenswrapper[4667]: I0929 17:20:58.902607 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" podStartSLOduration=37.291513935 podStartE2EDuration="38.902588827s" podCreationTimestamp="2025-09-29 17:20:20 +0000 UTC" firstStartedPulling="2025-09-29 17:20:57.154935566 +0000 UTC m=+685.652782335" lastFinishedPulling="2025-09-29 17:20:58.766010457 +0000 UTC m=+687.263857227" observedRunningTime="2025-09-29 17:20:58.899019757 +0000 UTC m=+687.396866526" watchObservedRunningTime="2025-09-29 17:20:58.902588827 +0000 UTC m=+687.400435596" Sep 29 17:20:59 crc kubenswrapper[4667]: I0929 17:20:59.882547 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" event={"ID":"04f22be1-6686-49ab-8abd-4d25428e1a8d","Type":"ContainerStarted","Data":"36e4b73f3bc36b1b92b6bdf3cfc7a6513636d88c81b089c1c99518227ef045f8"} Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.697947 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qh727"] Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.698676 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh727" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.701099 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.701326 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-tx8qq"] Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.701938 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.702095 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-tx8qq" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.702592 4667 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-97mst" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.705923 4667 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-j5zlh" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.712958 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-tx8qq"] Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.715933 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qh727"] Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.721118 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-bzmsv"] Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.723694 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.727522 4667 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-6ghr2" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.736018 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-bzmsv"] Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.786964 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs8ts\" (UniqueName: \"kubernetes.io/projected/1eb84072-ac55-4869-a958-6b2419c839ae-kube-api-access-xs8ts\") pod \"cert-manager-webhook-5655c58dd6-bzmsv\" (UID: \"1eb84072-ac55-4869-a958-6b2419c839ae\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.787011 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c245j\" (UniqueName: \"kubernetes.io/projected/77ca411f-8a0c-40b8-8457-57ed07672d2c-kube-api-access-c245j\") pod \"cert-manager-cainjector-7f985d654d-qh727\" (UID: \"77ca411f-8a0c-40b8-8457-57ed07672d2c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qh727" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.787043 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f4m5\" (UniqueName: \"kubernetes.io/projected/ce99b74e-e34b-4fbb-abfd-439f1c7f3585-kube-api-access-9f4m5\") pod \"cert-manager-5b446d88c5-tx8qq\" (UID: \"ce99b74e-e34b-4fbb-abfd-439f1c7f3585\") " pod="cert-manager/cert-manager-5b446d88c5-tx8qq" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.887987 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs8ts\" (UniqueName: \"kubernetes.io/projected/1eb84072-ac55-4869-a958-6b2419c839ae-kube-api-access-xs8ts\") pod \"cert-manager-webhook-5655c58dd6-bzmsv\" (UID: \"1eb84072-ac55-4869-a958-6b2419c839ae\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.888058 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c245j\" (UniqueName: \"kubernetes.io/projected/77ca411f-8a0c-40b8-8457-57ed07672d2c-kube-api-access-c245j\") pod \"cert-manager-cainjector-7f985d654d-qh727\" (UID: \"77ca411f-8a0c-40b8-8457-57ed07672d2c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qh727" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.888140 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f4m5\" (UniqueName: \"kubernetes.io/projected/ce99b74e-e34b-4fbb-abfd-439f1c7f3585-kube-api-access-9f4m5\") pod \"cert-manager-5b446d88c5-tx8qq\" (UID: \"ce99b74e-e34b-4fbb-abfd-439f1c7f3585\") " pod="cert-manager/cert-manager-5b446d88c5-tx8qq" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.894752 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" event={"ID":"04f22be1-6686-49ab-8abd-4d25428e1a8d","Type":"ContainerStarted","Data":"90370806c309e0de18100115529ebbb1c8cf2bb7adab57775876e7bef69b6a5d"} Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.903403 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c245j\" (UniqueName: \"kubernetes.io/projected/77ca411f-8a0c-40b8-8457-57ed07672d2c-kube-api-access-c245j\") pod \"cert-manager-cainjector-7f985d654d-qh727\" (UID: \"77ca411f-8a0c-40b8-8457-57ed07672d2c\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-qh727" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.904015 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f4m5\" (UniqueName: \"kubernetes.io/projected/ce99b74e-e34b-4fbb-abfd-439f1c7f3585-kube-api-access-9f4m5\") pod \"cert-manager-5b446d88c5-tx8qq\" (UID: \"ce99b74e-e34b-4fbb-abfd-439f1c7f3585\") " pod="cert-manager/cert-manager-5b446d88c5-tx8qq" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.905643 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs8ts\" (UniqueName: \"kubernetes.io/projected/1eb84072-ac55-4869-a958-6b2419c839ae-kube-api-access-xs8ts\") pod \"cert-manager-webhook-5655c58dd6-bzmsv\" (UID: \"1eb84072-ac55-4869-a958-6b2419c839ae\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" Sep 29 17:21:01 crc kubenswrapper[4667]: I0929 17:21:01.912386 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-x5wg4" podStartSLOduration=39.539167027 podStartE2EDuration="41.912375411s" podCreationTimestamp="2025-09-29 17:20:20 +0000 UTC" firstStartedPulling="2025-09-29 17:20:59.175629285 +0000 UTC m=+687.673476055" lastFinishedPulling="2025-09-29 17:21:01.54883767 +0000 UTC m=+690.046684439" observedRunningTime="2025-09-29 17:21:01.905291033 +0000 UTC m=+690.403137802" watchObservedRunningTime="2025-09-29 17:21:01.912375411 +0000 UTC m=+690.410222180" Sep 29 17:21:02 crc kubenswrapper[4667]: I0929 17:21:02.014464 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh727" Sep 29 17:21:02 crc kubenswrapper[4667]: I0929 17:21:02.023535 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-tx8qq" Sep 29 17:21:02 crc kubenswrapper[4667]: I0929 17:21:02.035588 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" Sep 29 17:21:02 crc kubenswrapper[4667]: I0929 17:21:02.387744 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-tx8qq"] Sep 29 17:21:02 crc kubenswrapper[4667]: W0929 17:21:02.395001 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce99b74e_e34b_4fbb_abfd_439f1c7f3585.slice/crio-e0266fb79f4a0fd834e31fb4aa5f1d38eb8433e85013ddb940b3877bd08683e2 WatchSource:0}: Error finding container e0266fb79f4a0fd834e31fb4aa5f1d38eb8433e85013ddb940b3877bd08683e2: Status 404 returned error can't find the container with id e0266fb79f4a0fd834e31fb4aa5f1d38eb8433e85013ddb940b3877bd08683e2 Sep 29 17:21:02 crc kubenswrapper[4667]: I0929 17:21:02.400244 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-bzmsv"] Sep 29 17:21:02 crc kubenswrapper[4667]: I0929 17:21:02.403272 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-qh727"] Sep 29 17:21:02 crc kubenswrapper[4667]: W0929 17:21:02.407798 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ca411f_8a0c_40b8_8457_57ed07672d2c.slice/crio-db044c7dd5a8cad4e29a71716deddff7e41e8c9285756b34f364f8100eb23169 WatchSource:0}: Error finding container db044c7dd5a8cad4e29a71716deddff7e41e8c9285756b34f364f8100eb23169: Status 404 returned error can't find the container with id db044c7dd5a8cad4e29a71716deddff7e41e8c9285756b34f364f8100eb23169 Sep 29 17:21:02 crc kubenswrapper[4667]: I0929 17:21:02.900545 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" event={"ID":"1eb84072-ac55-4869-a958-6b2419c839ae","Type":"ContainerStarted","Data":"51969ce15021958fcceb5b8007e41f4155658f99775712ce34d9c9315c2cce41"} Sep 29 17:21:02 crc kubenswrapper[4667]: I0929 17:21:02.901352 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh727" event={"ID":"77ca411f-8a0c-40b8-8457-57ed07672d2c","Type":"ContainerStarted","Data":"db044c7dd5a8cad4e29a71716deddff7e41e8c9285756b34f364f8100eb23169"} Sep 29 17:21:02 crc kubenswrapper[4667]: I0929 17:21:02.902534 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-tx8qq" event={"ID":"ce99b74e-e34b-4fbb-abfd-439f1c7f3585","Type":"ContainerStarted","Data":"e0266fb79f4a0fd834e31fb4aa5f1d38eb8433e85013ddb940b3877bd08683e2"} Sep 29 17:21:04 crc kubenswrapper[4667]: I0929 17:21:04.915255 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-tx8qq" event={"ID":"ce99b74e-e34b-4fbb-abfd-439f1c7f3585","Type":"ContainerStarted","Data":"c4d397b2df46fd12590ea844d269e3ad104470f2924317e959ffce5ca2178032"} Sep 29 17:21:04 crc kubenswrapper[4667]: I0929 17:21:04.917000 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh727" event={"ID":"77ca411f-8a0c-40b8-8457-57ed07672d2c","Type":"ContainerStarted","Data":"41cefe5da31dbf731a743eab29b62293ce797dec9c0d8b0a858b263f00484c0e"} Sep 29 17:21:04 crc kubenswrapper[4667]: I0929 17:21:04.939568 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-tx8qq" podStartSLOduration=1.9102457290000001 podStartE2EDuration="3.93955253s" podCreationTimestamp="2025-09-29 17:21:01 +0000 UTC" firstStartedPulling="2025-09-29 17:21:02.397296209 +0000 UTC m=+690.895142977" lastFinishedPulling="2025-09-29 17:21:04.426603009 +0000 UTC m=+692.924449778" observedRunningTime="2025-09-29 17:21:04.934817651 +0000 UTC m=+693.432664420" watchObservedRunningTime="2025-09-29 17:21:04.93955253 +0000 UTC m=+693.437399300" Sep 29 17:21:05 crc kubenswrapper[4667]: I0929 17:21:05.923205 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" event={"ID":"1eb84072-ac55-4869-a958-6b2419c839ae","Type":"ContainerStarted","Data":"ed5edbce905beff36abe6b7b27e3404ee55fea9e4003e5eb5dcd7e21bddcab3b"} Sep 29 17:21:05 crc kubenswrapper[4667]: I0929 17:21:05.937503 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-qh727" podStartSLOduration=2.891527946 podStartE2EDuration="4.93747705s" podCreationTimestamp="2025-09-29 17:21:01 +0000 UTC" firstStartedPulling="2025-09-29 17:21:02.409327777 +0000 UTC m=+690.907174545" lastFinishedPulling="2025-09-29 17:21:04.45527688 +0000 UTC m=+692.953123649" observedRunningTime="2025-09-29 17:21:04.957078489 +0000 UTC m=+693.454925257" watchObservedRunningTime="2025-09-29 17:21:05.93747705 +0000 UTC m=+694.435323818" Sep 29 17:21:05 crc kubenswrapper[4667]: I0929 17:21:05.938355 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" podStartSLOduration=2.077646711 podStartE2EDuration="4.938350057s" podCreationTimestamp="2025-09-29 17:21:01 +0000 UTC" firstStartedPulling="2025-09-29 17:21:02.403794102 +0000 UTC m=+690.901640870" lastFinishedPulling="2025-09-29 17:21:05.264497447 +0000 UTC m=+693.762344216" observedRunningTime="2025-09-29 17:21:05.934595155 +0000 UTC m=+694.432441925" watchObservedRunningTime="2025-09-29 17:21:05.938350057 +0000 UTC m=+694.436196825" Sep 29 17:21:06 crc kubenswrapper[4667]: I0929 17:21:06.931239 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" Sep 29 17:21:11 crc kubenswrapper[4667]: I0929 17:21:11.358515 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-rphbr" Sep 29 17:21:12 crc kubenswrapper[4667]: I0929 17:21:12.038291 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-bzmsv" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.236974 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb"] Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.238230 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.240176 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.251494 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb"] Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.311242 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss59p\" (UniqueName: \"kubernetes.io/projected/7f8a4805-c531-4623-a914-f76593c170ae-kube-api-access-ss59p\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.311289 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-bundle\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.311326 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-util\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.412933 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss59p\" (UniqueName: \"kubernetes.io/projected/7f8a4805-c531-4623-a914-f76593c170ae-kube-api-access-ss59p\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.412981 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-bundle\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.413009 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-util\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.413442 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-util\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.413622 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-bundle\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.413648 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs"] Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.414904 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.419146 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs"] Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.428538 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss59p\" (UniqueName: \"kubernetes.io/projected/7f8a4805-c531-4623-a914-f76593c170ae-kube-api-access-ss59p\") pod \"c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.514436 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-util\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.514472 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm6r5\" (UniqueName: \"kubernetes.io/projected/f334231f-deeb-406a-930f-53c1fb6d7b86-kube-api-access-vm6r5\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.514575 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-bundle\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.549670 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.615929 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-util\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.616075 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm6r5\" (UniqueName: \"kubernetes.io/projected/f334231f-deeb-406a-930f-53c1fb6d7b86-kube-api-access-vm6r5\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.616135 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-bundle\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.616342 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-util\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.616437 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-bundle\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.631412 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm6r5\" (UniqueName: \"kubernetes.io/projected/f334231f-deeb-406a-930f-53c1fb6d7b86-kube-api-access-vm6r5\") pod \"e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.725586 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:30 crc kubenswrapper[4667]: I0929 17:21:30.899023 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb"] Sep 29 17:21:31 crc kubenswrapper[4667]: I0929 17:21:31.042246 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" event={"ID":"7f8a4805-c531-4623-a914-f76593c170ae","Type":"ContainerStarted","Data":"2ba88b1b3320dd1331d1f4c36a44d755717ad8b4fe3761ce7b5dc84d2e9ecf0c"} Sep 29 17:21:31 crc kubenswrapper[4667]: I0929 17:21:31.054561 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs"] Sep 29 17:21:31 crc kubenswrapper[4667]: W0929 17:21:31.058209 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf334231f_deeb_406a_930f_53c1fb6d7b86.slice/crio-ce3b9b713b33e68dd4fd4f8df5033922d8acd7962b52cdc0d29697658b691a33 WatchSource:0}: Error finding container ce3b9b713b33e68dd4fd4f8df5033922d8acd7962b52cdc0d29697658b691a33: Status 404 returned error can't find the container with id ce3b9b713b33e68dd4fd4f8df5033922d8acd7962b52cdc0d29697658b691a33 Sep 29 17:21:32 crc kubenswrapper[4667]: I0929 17:21:32.046367 4667 generic.go:334] "Generic (PLEG): container finished" podID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerID="c61a3020c9c59e756e46d64d4809fc4235b8f01a76e3036601c7500c51fe748d" exitCode=0 Sep 29 17:21:32 crc kubenswrapper[4667]: I0929 17:21:32.046396 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" event={"ID":"f334231f-deeb-406a-930f-53c1fb6d7b86","Type":"ContainerDied","Data":"c61a3020c9c59e756e46d64d4809fc4235b8f01a76e3036601c7500c51fe748d"} Sep 29 17:21:32 crc kubenswrapper[4667]: I0929 17:21:32.046610 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" event={"ID":"f334231f-deeb-406a-930f-53c1fb6d7b86","Type":"ContainerStarted","Data":"ce3b9b713b33e68dd4fd4f8df5033922d8acd7962b52cdc0d29697658b691a33"} Sep 29 17:21:32 crc kubenswrapper[4667]: I0929 17:21:32.047765 4667 generic.go:334] "Generic (PLEG): container finished" podID="7f8a4805-c531-4623-a914-f76593c170ae" containerID="9b743ccbbce1bcf5d9b1732cebccf4cc89f1105ad2d73aa92eaa37ab3cdb495c" exitCode=0 Sep 29 17:21:32 crc kubenswrapper[4667]: I0929 17:21:32.047797 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" event={"ID":"7f8a4805-c531-4623-a914-f76593c170ae","Type":"ContainerDied","Data":"9b743ccbbce1bcf5d9b1732cebccf4cc89f1105ad2d73aa92eaa37ab3cdb495c"} Sep 29 17:21:35 crc kubenswrapper[4667]: I0929 17:21:35.062086 4667 generic.go:334] "Generic (PLEG): container finished" podID="7f8a4805-c531-4623-a914-f76593c170ae" containerID="08fdc6f8e8a1b71390feceb5b961593d8b31fd11eb9be0ccfa99e544fd6fdd09" exitCode=0 Sep 29 17:21:35 crc kubenswrapper[4667]: I0929 17:21:35.062123 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" event={"ID":"7f8a4805-c531-4623-a914-f76593c170ae","Type":"ContainerDied","Data":"08fdc6f8e8a1b71390feceb5b961593d8b31fd11eb9be0ccfa99e544fd6fdd09"} Sep 29 17:21:35 crc kubenswrapper[4667]: I0929 17:21:35.063999 4667 generic.go:334] "Generic (PLEG): container finished" podID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerID="49ffb3eef658eeafb1f27778429bb66824f7488a08c154007b79e27536b73e52" exitCode=0 Sep 29 17:21:35 crc kubenswrapper[4667]: I0929 17:21:35.064034 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" event={"ID":"f334231f-deeb-406a-930f-53c1fb6d7b86","Type":"ContainerDied","Data":"49ffb3eef658eeafb1f27778429bb66824f7488a08c154007b79e27536b73e52"} Sep 29 17:21:36 crc kubenswrapper[4667]: I0929 17:21:36.071043 4667 generic.go:334] "Generic (PLEG): container finished" podID="7f8a4805-c531-4623-a914-f76593c170ae" containerID="832f6b2636a2d13b47bf4a6d8fe0b90ef1f8dc7cb397f9ce3620787b0c52d8a2" exitCode=0 Sep 29 17:21:36 crc kubenswrapper[4667]: I0929 17:21:36.071082 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" event={"ID":"7f8a4805-c531-4623-a914-f76593c170ae","Type":"ContainerDied","Data":"832f6b2636a2d13b47bf4a6d8fe0b90ef1f8dc7cb397f9ce3620787b0c52d8a2"} Sep 29 17:21:36 crc kubenswrapper[4667]: I0929 17:21:36.072997 4667 generic.go:334] "Generic (PLEG): container finished" podID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerID="6cf239f14df7cc61f29d4b2570480669919c34c162a967bab0b14434d0fc74e7" exitCode=0 Sep 29 17:21:36 crc kubenswrapper[4667]: I0929 17:21:36.073035 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" event={"ID":"f334231f-deeb-406a-930f-53c1fb6d7b86","Type":"ContainerDied","Data":"6cf239f14df7cc61f29d4b2570480669919c34c162a967bab0b14434d0fc74e7"} Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.277097 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.280167 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.291581 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-util\") pod \"f334231f-deeb-406a-930f-53c1fb6d7b86\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.300456 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-util" (OuterVolumeSpecName: "util") pod "f334231f-deeb-406a-930f-53c1fb6d7b86" (UID: "f334231f-deeb-406a-930f-53c1fb6d7b86"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.392787 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-util\") pod \"7f8a4805-c531-4623-a914-f76593c170ae\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.393823 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-bundle\") pod \"7f8a4805-c531-4623-a914-f76593c170ae\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.393900 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss59p\" (UniqueName: \"kubernetes.io/projected/7f8a4805-c531-4623-a914-f76593c170ae-kube-api-access-ss59p\") pod \"7f8a4805-c531-4623-a914-f76593c170ae\" (UID: \"7f8a4805-c531-4623-a914-f76593c170ae\") " Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.393932 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-bundle\") pod \"f334231f-deeb-406a-930f-53c1fb6d7b86\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.393947 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vm6r5\" (UniqueName: \"kubernetes.io/projected/f334231f-deeb-406a-930f-53c1fb6d7b86-kube-api-access-vm6r5\") pod \"f334231f-deeb-406a-930f-53c1fb6d7b86\" (UID: \"f334231f-deeb-406a-930f-53c1fb6d7b86\") " Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.394211 4667 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-util\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.395212 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-bundle" (OuterVolumeSpecName: "bundle") pod "f334231f-deeb-406a-930f-53c1fb6d7b86" (UID: "f334231f-deeb-406a-930f-53c1fb6d7b86"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.395408 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-bundle" (OuterVolumeSpecName: "bundle") pod "7f8a4805-c531-4623-a914-f76593c170ae" (UID: "7f8a4805-c531-4623-a914-f76593c170ae"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.398385 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f334231f-deeb-406a-930f-53c1fb6d7b86-kube-api-access-vm6r5" (OuterVolumeSpecName: "kube-api-access-vm6r5") pod "f334231f-deeb-406a-930f-53c1fb6d7b86" (UID: "f334231f-deeb-406a-930f-53c1fb6d7b86"). InnerVolumeSpecName "kube-api-access-vm6r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.398973 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f8a4805-c531-4623-a914-f76593c170ae-kube-api-access-ss59p" (OuterVolumeSpecName: "kube-api-access-ss59p") pod "7f8a4805-c531-4623-a914-f76593c170ae" (UID: "7f8a4805-c531-4623-a914-f76593c170ae"). InnerVolumeSpecName "kube-api-access-ss59p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.401478 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-util" (OuterVolumeSpecName: "util") pod "7f8a4805-c531-4623-a914-f76593c170ae" (UID: "7f8a4805-c531-4623-a914-f76593c170ae"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.494740 4667 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-util\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.494774 4667 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f8a4805-c531-4623-a914-f76593c170ae-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.494784 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss59p\" (UniqueName: \"kubernetes.io/projected/7f8a4805-c531-4623-a914-f76593c170ae-kube-api-access-ss59p\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.494794 4667 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f334231f-deeb-406a-930f-53c1fb6d7b86-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:37 crc kubenswrapper[4667]: I0929 17:21:37.494803 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vm6r5\" (UniqueName: \"kubernetes.io/projected/f334231f-deeb-406a-930f-53c1fb6d7b86-kube-api-access-vm6r5\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:38 crc kubenswrapper[4667]: I0929 17:21:38.083505 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" event={"ID":"f334231f-deeb-406a-930f-53c1fb6d7b86","Type":"ContainerDied","Data":"ce3b9b713b33e68dd4fd4f8df5033922d8acd7962b52cdc0d29697658b691a33"} Sep 29 17:21:38 crc kubenswrapper[4667]: I0929 17:21:38.083551 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce3b9b713b33e68dd4fd4f8df5033922d8acd7962b52cdc0d29697658b691a33" Sep 29 17:21:38 crc kubenswrapper[4667]: I0929 17:21:38.083522 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs" Sep 29 17:21:38 crc kubenswrapper[4667]: I0929 17:21:38.086801 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" event={"ID":"7f8a4805-c531-4623-a914-f76593c170ae","Type":"ContainerDied","Data":"2ba88b1b3320dd1331d1f4c36a44d755717ad8b4fe3761ce7b5dc84d2e9ecf0c"} Sep 29 17:21:38 crc kubenswrapper[4667]: I0929 17:21:38.086827 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb" Sep 29 17:21:38 crc kubenswrapper[4667]: I0929 17:21:38.086835 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ba88b1b3320dd1331d1f4c36a44d755717ad8b4fe3761ce7b5dc84d2e9ecf0c" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.585293 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd"] Sep 29 17:21:48 crc kubenswrapper[4667]: E0929 17:21:48.585941 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f8a4805-c531-4623-a914-f76593c170ae" containerName="pull" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.585953 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f8a4805-c531-4623-a914-f76593c170ae" containerName="pull" Sep 29 17:21:48 crc kubenswrapper[4667]: E0929 17:21:48.585965 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerName="pull" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.585970 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerName="pull" Sep 29 17:21:48 crc kubenswrapper[4667]: E0929 17:21:48.585979 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerName="util" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.585984 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerName="util" Sep 29 17:21:48 crc kubenswrapper[4667]: E0929 17:21:48.585992 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerName="extract" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.585997 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerName="extract" Sep 29 17:21:48 crc kubenswrapper[4667]: E0929 17:21:48.586012 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f8a4805-c531-4623-a914-f76593c170ae" containerName="extract" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.586017 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f8a4805-c531-4623-a914-f76593c170ae" containerName="extract" Sep 29 17:21:48 crc kubenswrapper[4667]: E0929 17:21:48.586025 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f8a4805-c531-4623-a914-f76593c170ae" containerName="util" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.586030 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f8a4805-c531-4623-a914-f76593c170ae" containerName="util" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.586116 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="f334231f-deeb-406a-930f-53c1fb6d7b86" containerName="extract" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.586131 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f8a4805-c531-4623-a914-f76593c170ae" containerName="extract" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.586622 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.588374 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.588408 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.588580 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.588618 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.588780 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.589510 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-kl7zv" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.609797 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd"] Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.721808 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbde58f9-2869-40cd-bef8-a28336337936-webhook-cert\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.721912 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/fbde58f9-2869-40cd-bef8-a28336337936-manager-config\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.721948 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbde58f9-2869-40cd-bef8-a28336337936-apiservice-cert\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.722006 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fbde58f9-2869-40cd-bef8-a28336337936-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.722030 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmc5m\" (UniqueName: \"kubernetes.io/projected/fbde58f9-2869-40cd-bef8-a28336337936-kube-api-access-kmc5m\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.823372 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/fbde58f9-2869-40cd-bef8-a28336337936-manager-config\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.823652 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbde58f9-2869-40cd-bef8-a28336337936-apiservice-cert\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.823724 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fbde58f9-2869-40cd-bef8-a28336337936-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.823756 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmc5m\" (UniqueName: \"kubernetes.io/projected/fbde58f9-2869-40cd-bef8-a28336337936-kube-api-access-kmc5m\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.823916 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbde58f9-2869-40cd-bef8-a28336337936-webhook-cert\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.824218 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/fbde58f9-2869-40cd-bef8-a28336337936-manager-config\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.828865 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fbde58f9-2869-40cd-bef8-a28336337936-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.828918 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fbde58f9-2869-40cd-bef8-a28336337936-webhook-cert\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.833289 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fbde58f9-2869-40cd-bef8-a28336337936-apiservice-cert\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.839048 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmc5m\" (UniqueName: \"kubernetes.io/projected/fbde58f9-2869-40cd-bef8-a28336337936-kube-api-access-kmc5m\") pod \"loki-operator-controller-manager-679b68c759-tgpzd\" (UID: \"fbde58f9-2869-40cd-bef8-a28336337936\") " pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:48 crc kubenswrapper[4667]: I0929 17:21:48.903379 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.292435 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd"] Sep 29 17:21:49 crc kubenswrapper[4667]: W0929 17:21:49.298271 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbde58f9_2869_40cd_bef8_a28336337936.slice/crio-68eecadcb2a42f6ab0d7e3ab8f731e4ce9fe127debacf6c97e3acbdf76e9d78a WatchSource:0}: Error finding container 68eecadcb2a42f6ab0d7e3ab8f731e4ce9fe127debacf6c97e3acbdf76e9d78a: Status 404 returned error can't find the container with id 68eecadcb2a42f6ab0d7e3ab8f731e4ce9fe127debacf6c97e3acbdf76e9d78a Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.345814 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-fcc886d58-crc96"] Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.346458 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-fcc886d58-crc96" Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.350117 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.352048 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-5t4bf" Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.352152 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.363822 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-fcc886d58-crc96"] Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.432550 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c22kf\" (UniqueName: \"kubernetes.io/projected/3519c647-b665-4a91-bcf2-3abbd6f982c9-kube-api-access-c22kf\") pod \"cluster-logging-operator-fcc886d58-crc96\" (UID: \"3519c647-b665-4a91-bcf2-3abbd6f982c9\") " pod="openshift-logging/cluster-logging-operator-fcc886d58-crc96" Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.534036 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c22kf\" (UniqueName: \"kubernetes.io/projected/3519c647-b665-4a91-bcf2-3abbd6f982c9-kube-api-access-c22kf\") pod \"cluster-logging-operator-fcc886d58-crc96\" (UID: \"3519c647-b665-4a91-bcf2-3abbd6f982c9\") " pod="openshift-logging/cluster-logging-operator-fcc886d58-crc96" Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.552308 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c22kf\" (UniqueName: \"kubernetes.io/projected/3519c647-b665-4a91-bcf2-3abbd6f982c9-kube-api-access-c22kf\") pod \"cluster-logging-operator-fcc886d58-crc96\" (UID: \"3519c647-b665-4a91-bcf2-3abbd6f982c9\") " pod="openshift-logging/cluster-logging-operator-fcc886d58-crc96" Sep 29 17:21:49 crc kubenswrapper[4667]: I0929 17:21:49.663387 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-fcc886d58-crc96" Sep 29 17:21:50 crc kubenswrapper[4667]: I0929 17:21:50.046317 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-fcc886d58-crc96"] Sep 29 17:21:50 crc kubenswrapper[4667]: W0929 17:21:50.050732 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3519c647_b665_4a91_bcf2_3abbd6f982c9.slice/crio-5492944a23586ac50951f203304e6d6a0168c2b90104747eae8114ed018ece77 WatchSource:0}: Error finding container 5492944a23586ac50951f203304e6d6a0168c2b90104747eae8114ed018ece77: Status 404 returned error can't find the container with id 5492944a23586ac50951f203304e6d6a0168c2b90104747eae8114ed018ece77 Sep 29 17:21:50 crc kubenswrapper[4667]: I0929 17:21:50.142941 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" event={"ID":"fbde58f9-2869-40cd-bef8-a28336337936","Type":"ContainerStarted","Data":"68eecadcb2a42f6ab0d7e3ab8f731e4ce9fe127debacf6c97e3acbdf76e9d78a"} Sep 29 17:21:50 crc kubenswrapper[4667]: I0929 17:21:50.144393 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-fcc886d58-crc96" event={"ID":"3519c647-b665-4a91-bcf2-3abbd6f982c9","Type":"ContainerStarted","Data":"5492944a23586ac50951f203304e6d6a0168c2b90104747eae8114ed018ece77"} Sep 29 17:21:51 crc kubenswrapper[4667]: I0929 17:21:51.579505 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5l58z"] Sep 29 17:21:51 crc kubenswrapper[4667]: I0929 17:21:51.579957 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" podUID="9807d262-d148-47f0-afd5-9b92aacfba38" containerName="controller-manager" containerID="cri-o://3fe7830a7842f435ab64ef2b7fcfe4eb83ba8422c286f56fc90c974eea0c4126" gracePeriod=30 Sep 29 17:21:51 crc kubenswrapper[4667]: I0929 17:21:51.636855 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp"] Sep 29 17:21:51 crc kubenswrapper[4667]: I0929 17:21:51.637131 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" podUID="80a6113d-cddb-48e4-9855-73314af480f7" containerName="route-controller-manager" containerID="cri-o://4d3a7da769e75c5acbff30646b60d96418e90162cd4fe8acfa20bbaba188d5b9" gracePeriod=30 Sep 29 17:21:52 crc kubenswrapper[4667]: I0929 17:21:52.194464 4667 generic.go:334] "Generic (PLEG): container finished" podID="9807d262-d148-47f0-afd5-9b92aacfba38" containerID="3fe7830a7842f435ab64ef2b7fcfe4eb83ba8422c286f56fc90c974eea0c4126" exitCode=0 Sep 29 17:21:52 crc kubenswrapper[4667]: I0929 17:21:52.194552 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" event={"ID":"9807d262-d148-47f0-afd5-9b92aacfba38","Type":"ContainerDied","Data":"3fe7830a7842f435ab64ef2b7fcfe4eb83ba8422c286f56fc90c974eea0c4126"} Sep 29 17:21:52 crc kubenswrapper[4667]: I0929 17:21:52.202599 4667 generic.go:334] "Generic (PLEG): container finished" podID="80a6113d-cddb-48e4-9855-73314af480f7" containerID="4d3a7da769e75c5acbff30646b60d96418e90162cd4fe8acfa20bbaba188d5b9" exitCode=0 Sep 29 17:21:52 crc kubenswrapper[4667]: I0929 17:21:52.202644 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" event={"ID":"80a6113d-cddb-48e4-9855-73314af480f7","Type":"ContainerDied","Data":"4d3a7da769e75c5acbff30646b60d96418e90162cd4fe8acfa20bbaba188d5b9"} Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.386339 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.402398 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.415084 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh"] Sep 29 17:21:53 crc kubenswrapper[4667]: E0929 17:21:53.415302 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a6113d-cddb-48e4-9855-73314af480f7" containerName="route-controller-manager" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.415319 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a6113d-cddb-48e4-9855-73314af480f7" containerName="route-controller-manager" Sep 29 17:21:53 crc kubenswrapper[4667]: E0929 17:21:53.415332 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9807d262-d148-47f0-afd5-9b92aacfba38" containerName="controller-manager" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.415337 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="9807d262-d148-47f0-afd5-9b92aacfba38" containerName="controller-manager" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.415425 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="9807d262-d148-47f0-afd5-9b92aacfba38" containerName="controller-manager" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.415438 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="80a6113d-cddb-48e4-9855-73314af480f7" containerName="route-controller-manager" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.415784 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.426910 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh"] Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498204 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xhc\" (UniqueName: \"kubernetes.io/projected/9807d262-d148-47f0-afd5-9b92aacfba38-kube-api-access-w4xhc\") pod \"9807d262-d148-47f0-afd5-9b92aacfba38\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498249 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a6113d-cddb-48e4-9855-73314af480f7-serving-cert\") pod \"80a6113d-cddb-48e4-9855-73314af480f7\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498281 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-client-ca\") pod \"80a6113d-cddb-48e4-9855-73314af480f7\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498302 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-config\") pod \"80a6113d-cddb-48e4-9855-73314af480f7\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498329 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-config\") pod \"9807d262-d148-47f0-afd5-9b92aacfba38\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498348 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9807d262-d148-47f0-afd5-9b92aacfba38-serving-cert\") pod \"9807d262-d148-47f0-afd5-9b92aacfba38\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498379 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-proxy-ca-bundles\") pod \"9807d262-d148-47f0-afd5-9b92aacfba38\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498398 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrdrs\" (UniqueName: \"kubernetes.io/projected/80a6113d-cddb-48e4-9855-73314af480f7-kube-api-access-jrdrs\") pod \"80a6113d-cddb-48e4-9855-73314af480f7\" (UID: \"80a6113d-cddb-48e4-9855-73314af480f7\") " Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498412 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-client-ca\") pod \"9807d262-d148-47f0-afd5-9b92aacfba38\" (UID: \"9807d262-d148-47f0-afd5-9b92aacfba38\") " Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498537 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-config\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498561 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-serving-cert\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498579 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvtk9\" (UniqueName: \"kubernetes.io/projected/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-kube-api-access-nvtk9\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498614 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-client-ca\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.498638 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-proxy-ca-bundles\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.500168 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9807d262-d148-47f0-afd5-9b92aacfba38" (UID: "9807d262-d148-47f0-afd5-9b92aacfba38"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.500246 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-config" (OuterVolumeSpecName: "config") pod "80a6113d-cddb-48e4-9855-73314af480f7" (UID: "80a6113d-cddb-48e4-9855-73314af480f7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.500676 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-config" (OuterVolumeSpecName: "config") pod "9807d262-d148-47f0-afd5-9b92aacfba38" (UID: "9807d262-d148-47f0-afd5-9b92aacfba38"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.506234 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9807d262-d148-47f0-afd5-9b92aacfba38-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9807d262-d148-47f0-afd5-9b92aacfba38" (UID: "9807d262-d148-47f0-afd5-9b92aacfba38"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.506234 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80a6113d-cddb-48e4-9855-73314af480f7-kube-api-access-jrdrs" (OuterVolumeSpecName: "kube-api-access-jrdrs") pod "80a6113d-cddb-48e4-9855-73314af480f7" (UID: "80a6113d-cddb-48e4-9855-73314af480f7"). InnerVolumeSpecName "kube-api-access-jrdrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.506551 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-client-ca" (OuterVolumeSpecName: "client-ca") pod "80a6113d-cddb-48e4-9855-73314af480f7" (UID: "80a6113d-cddb-48e4-9855-73314af480f7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.506683 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-client-ca" (OuterVolumeSpecName: "client-ca") pod "9807d262-d148-47f0-afd5-9b92aacfba38" (UID: "9807d262-d148-47f0-afd5-9b92aacfba38"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.508858 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80a6113d-cddb-48e4-9855-73314af480f7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "80a6113d-cddb-48e4-9855-73314af480f7" (UID: "80a6113d-cddb-48e4-9855-73314af480f7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.508920 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9807d262-d148-47f0-afd5-9b92aacfba38-kube-api-access-w4xhc" (OuterVolumeSpecName: "kube-api-access-w4xhc") pod "9807d262-d148-47f0-afd5-9b92aacfba38" (UID: "9807d262-d148-47f0-afd5-9b92aacfba38"). InnerVolumeSpecName "kube-api-access-w4xhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599333 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-config\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599390 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-serving-cert\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599421 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvtk9\" (UniqueName: \"kubernetes.io/projected/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-kube-api-access-nvtk9\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599478 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-client-ca\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599517 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-proxy-ca-bundles\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599639 4667 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599650 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80a6113d-cddb-48e4-9855-73314af480f7-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599658 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599665 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9807d262-d148-47f0-afd5-9b92aacfba38-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599675 4667 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599683 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrdrs\" (UniqueName: \"kubernetes.io/projected/80a6113d-cddb-48e4-9855-73314af480f7-kube-api-access-jrdrs\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599691 4667 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9807d262-d148-47f0-afd5-9b92aacfba38-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599721 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xhc\" (UniqueName: \"kubernetes.io/projected/9807d262-d148-47f0-afd5-9b92aacfba38-kube-api-access-w4xhc\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.599730 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/80a6113d-cddb-48e4-9855-73314af480f7-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.601207 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-proxy-ca-bundles\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.601367 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-config\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.601643 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-client-ca\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.609679 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-serving-cert\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.627018 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvtk9\" (UniqueName: \"kubernetes.io/projected/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-kube-api-access-nvtk9\") pod \"controller-manager-5bdcf5fb77-qg4kh\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.662498 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh"] Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.663437 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.696570 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk"] Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.697257 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.707174 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk"] Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.801775 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/299789aa-c363-47e8-8a61-8eead4e082c5-serving-cert\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.801817 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/299789aa-c363-47e8-8a61-8eead4e082c5-client-ca\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.801858 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hnv6\" (UniqueName: \"kubernetes.io/projected/299789aa-c363-47e8-8a61-8eead4e082c5-kube-api-access-5hnv6\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.801877 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/299789aa-c363-47e8-8a61-8eead4e082c5-config\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.903452 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/299789aa-c363-47e8-8a61-8eead4e082c5-serving-cert\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.904956 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/299789aa-c363-47e8-8a61-8eead4e082c5-client-ca\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.906801 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/299789aa-c363-47e8-8a61-8eead4e082c5-client-ca\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.906893 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hnv6\" (UniqueName: \"kubernetes.io/projected/299789aa-c363-47e8-8a61-8eead4e082c5-kube-api-access-5hnv6\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.906922 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/299789aa-c363-47e8-8a61-8eead4e082c5-config\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.907870 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/299789aa-c363-47e8-8a61-8eead4e082c5-serving-cert\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.909640 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/299789aa-c363-47e8-8a61-8eead4e082c5-config\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:53 crc kubenswrapper[4667]: I0929 17:21:53.926986 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hnv6\" (UniqueName: \"kubernetes.io/projected/299789aa-c363-47e8-8a61-8eead4e082c5-kube-api-access-5hnv6\") pod \"route-controller-manager-6d96fdb6d5-d2dnk\" (UID: \"299789aa-c363-47e8-8a61-8eead4e082c5\") " pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.009344 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.217822 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" event={"ID":"80a6113d-cddb-48e4-9855-73314af480f7","Type":"ContainerDied","Data":"555d15c8e84cbbecc8f3f9d243d6dba46699d1072b2f61a39a4ac6bf8b47e60d"} Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.217859 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp" Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.217894 4667 scope.go:117] "RemoveContainer" containerID="4d3a7da769e75c5acbff30646b60d96418e90162cd4fe8acfa20bbaba188d5b9" Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.223005 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" event={"ID":"9807d262-d148-47f0-afd5-9b92aacfba38","Type":"ContainerDied","Data":"ef7e7eb075584f6603d9641ce5ce3feac4a4fa4781615bab0f7dce26bc3ecc5a"} Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.223050 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-5l58z" Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.237690 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp"] Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.241681 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-8qlvp"] Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.249957 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5l58z"] Sep 29 17:21:54 crc kubenswrapper[4667]: I0929 17:21:54.254273 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-5l58z"] Sep 29 17:21:55 crc kubenswrapper[4667]: I0929 17:21:55.440135 4667 scope.go:117] "RemoveContainer" containerID="3fe7830a7842f435ab64ef2b7fcfe4eb83ba8422c286f56fc90c974eea0c4126" Sep 29 17:21:55 crc kubenswrapper[4667]: I0929 17:21:55.714682 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk"] Sep 29 17:21:55 crc kubenswrapper[4667]: I0929 17:21:55.823734 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80a6113d-cddb-48e4-9855-73314af480f7" path="/var/lib/kubelet/pods/80a6113d-cddb-48e4-9855-73314af480f7/volumes" Sep 29 17:21:55 crc kubenswrapper[4667]: I0929 17:21:55.824883 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9807d262-d148-47f0-afd5-9b92aacfba38" path="/var/lib/kubelet/pods/9807d262-d148-47f0-afd5-9b92aacfba38/volumes" Sep 29 17:21:55 crc kubenswrapper[4667]: I0929 17:21:55.865209 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh"] Sep 29 17:21:55 crc kubenswrapper[4667]: W0929 17:21:55.872502 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b1169c8_5ea5_4382_8907_c2b5a7eab82c.slice/crio-17b8979a130535540dc4db460f7a754e1493950cfe7199d1a933f22d79172ba1 WatchSource:0}: Error finding container 17b8979a130535540dc4db460f7a754e1493950cfe7199d1a933f22d79172ba1: Status 404 returned error can't find the container with id 17b8979a130535540dc4db460f7a754e1493950cfe7199d1a933f22d79172ba1 Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.187660 4667 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.237258 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" event={"ID":"299789aa-c363-47e8-8a61-8eead4e082c5","Type":"ContainerStarted","Data":"b9a26cf95887dd3f62177f88cf26d44748e7242504df70bb3056ffb7438b7125"} Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.237529 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.237541 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" event={"ID":"299789aa-c363-47e8-8a61-8eead4e082c5","Type":"ContainerStarted","Data":"18d682debe36daf47fe253a3b2da4084278e00f60408c714d162a2eea1615194"} Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.239779 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" event={"ID":"fbde58f9-2869-40cd-bef8-a28336337936","Type":"ContainerStarted","Data":"425e8ecdd1eae85e2c5aa51f28f22241196f71b7d9276cce532ad56494c6d38d"} Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.241310 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" event={"ID":"9b1169c8-5ea5-4382-8907-c2b5a7eab82c","Type":"ContainerStarted","Data":"67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170"} Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.241339 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" event={"ID":"9b1169c8-5ea5-4382-8907-c2b5a7eab82c","Type":"ContainerStarted","Data":"17b8979a130535540dc4db460f7a754e1493950cfe7199d1a933f22d79172ba1"} Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.241422 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" podUID="9b1169c8-5ea5-4382-8907-c2b5a7eab82c" containerName="controller-manager" containerID="cri-o://67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170" gracePeriod=30 Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.241791 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.250927 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.257922 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" podStartSLOduration=3.25790996 podStartE2EDuration="3.25790996s" podCreationTimestamp="2025-09-29 17:21:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:21:56.256091523 +0000 UTC m=+744.753938291" watchObservedRunningTime="2025-09-29 17:21:56.25790996 +0000 UTC m=+744.755756730" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.261717 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-fcc886d58-crc96" event={"ID":"3519c647-b665-4a91-bcf2-3abbd6f982c9","Type":"ContainerStarted","Data":"660cfe04f9aa7f2ce8e5cc096ce0584f90c70e0e2543da63875f758d1b9516a3"} Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.277594 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d96fdb6d5-d2dnk" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.289367 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" podStartSLOduration=5.289353572 podStartE2EDuration="5.289353572s" podCreationTimestamp="2025-09-29 17:21:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:21:56.28549065 +0000 UTC m=+744.783337419" watchObservedRunningTime="2025-09-29 17:21:56.289353572 +0000 UTC m=+744.787200342" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.313128 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-fcc886d58-crc96" podStartSLOduration=1.8889081349999999 podStartE2EDuration="7.313112899s" podCreationTimestamp="2025-09-29 17:21:49 +0000 UTC" firstStartedPulling="2025-09-29 17:21:50.051992925 +0000 UTC m=+738.549839695" lastFinishedPulling="2025-09-29 17:21:55.47619769 +0000 UTC m=+743.974044459" observedRunningTime="2025-09-29 17:21:56.312354289 +0000 UTC m=+744.810201058" watchObservedRunningTime="2025-09-29 17:21:56.313112899 +0000 UTC m=+744.810959658" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.569546 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.588617 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66c94b6f9-lnxlc"] Sep 29 17:21:56 crc kubenswrapper[4667]: E0929 17:21:56.588825 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b1169c8-5ea5-4382-8907-c2b5a7eab82c" containerName="controller-manager" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.588859 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b1169c8-5ea5-4382-8907-c2b5a7eab82c" containerName="controller-manager" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.588966 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b1169c8-5ea5-4382-8907-c2b5a7eab82c" containerName="controller-manager" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.589331 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.604032 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66c94b6f9-lnxlc"] Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643098 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvtk9\" (UniqueName: \"kubernetes.io/projected/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-kube-api-access-nvtk9\") pod \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643165 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-client-ca\") pod \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643190 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-proxy-ca-bundles\") pod \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643250 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-serving-cert\") pod \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643345 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-config\") pod \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\" (UID: \"9b1169c8-5ea5-4382-8907-c2b5a7eab82c\") " Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643499 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-client-ca\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643540 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqn7v\" (UniqueName: \"kubernetes.io/projected/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-kube-api-access-jqn7v\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643559 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-serving-cert\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643575 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-config\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.643611 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-proxy-ca-bundles\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.644010 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-client-ca" (OuterVolumeSpecName: "client-ca") pod "9b1169c8-5ea5-4382-8907-c2b5a7eab82c" (UID: "9b1169c8-5ea5-4382-8907-c2b5a7eab82c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.644032 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-config" (OuterVolumeSpecName: "config") pod "9b1169c8-5ea5-4382-8907-c2b5a7eab82c" (UID: "9b1169c8-5ea5-4382-8907-c2b5a7eab82c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.644058 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9b1169c8-5ea5-4382-8907-c2b5a7eab82c" (UID: "9b1169c8-5ea5-4382-8907-c2b5a7eab82c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.648043 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9b1169c8-5ea5-4382-8907-c2b5a7eab82c" (UID: "9b1169c8-5ea5-4382-8907-c2b5a7eab82c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.648516 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-kube-api-access-nvtk9" (OuterVolumeSpecName: "kube-api-access-nvtk9") pod "9b1169c8-5ea5-4382-8907-c2b5a7eab82c" (UID: "9b1169c8-5ea5-4382-8907-c2b5a7eab82c"). InnerVolumeSpecName "kube-api-access-nvtk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745098 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-client-ca\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745166 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqn7v\" (UniqueName: \"kubernetes.io/projected/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-kube-api-access-jqn7v\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745187 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-serving-cert\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745203 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-config\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745230 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-proxy-ca-bundles\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745309 4667 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745321 4667 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745332 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvtk9\" (UniqueName: \"kubernetes.io/projected/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-kube-api-access-nvtk9\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745342 4667 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-client-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.745351 4667 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b1169c8-5ea5-4382-8907-c2b5a7eab82c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.746217 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-client-ca\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.746729 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-config\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.746786 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-proxy-ca-bundles\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.751740 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-serving-cert\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.760913 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqn7v\" (UniqueName: \"kubernetes.io/projected/7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462-kube-api-access-jqn7v\") pod \"controller-manager-66c94b6f9-lnxlc\" (UID: \"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462\") " pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:56 crc kubenswrapper[4667]: I0929 17:21:56.901906 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.269577 4667 generic.go:334] "Generic (PLEG): container finished" podID="9b1169c8-5ea5-4382-8907-c2b5a7eab82c" containerID="67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170" exitCode=0 Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.269717 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" event={"ID":"9b1169c8-5ea5-4382-8907-c2b5a7eab82c","Type":"ContainerDied","Data":"67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170"} Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.269743 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.270153 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh" event={"ID":"9b1169c8-5ea5-4382-8907-c2b5a7eab82c","Type":"ContainerDied","Data":"17b8979a130535540dc4db460f7a754e1493950cfe7199d1a933f22d79172ba1"} Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.270202 4667 scope.go:117] "RemoveContainer" containerID="67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170" Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.292902 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66c94b6f9-lnxlc"] Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.297588 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh"] Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.300254 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5bdcf5fb77-qg4kh"] Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.300620 4667 scope.go:117] "RemoveContainer" containerID="67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170" Sep 29 17:21:57 crc kubenswrapper[4667]: E0929 17:21:57.301185 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170\": container with ID starting with 67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170 not found: ID does not exist" containerID="67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170" Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.301227 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170"} err="failed to get container status \"67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170\": rpc error: code = NotFound desc = could not find container \"67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170\": container with ID starting with 67731e3ee7a664075a86ee64cf0da329ae6e941808268561116324d97c827170 not found: ID does not exist" Sep 29 17:21:57 crc kubenswrapper[4667]: W0929 17:21:57.307158 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7aba451b_e6ee_41e6_b5f5_ffbcbf9c5462.slice/crio-1b8d0ddbfad4e13c2bab623dbe836e996edcefd41f5cd2e1af120de5360e9f61 WatchSource:0}: Error finding container 1b8d0ddbfad4e13c2bab623dbe836e996edcefd41f5cd2e1af120de5360e9f61: Status 404 returned error can't find the container with id 1b8d0ddbfad4e13c2bab623dbe836e996edcefd41f5cd2e1af120de5360e9f61 Sep 29 17:21:57 crc kubenswrapper[4667]: I0929 17:21:57.824660 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b1169c8-5ea5-4382-8907-c2b5a7eab82c" path="/var/lib/kubelet/pods/9b1169c8-5ea5-4382-8907-c2b5a7eab82c/volumes" Sep 29 17:21:58 crc kubenswrapper[4667]: I0929 17:21:58.278252 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" event={"ID":"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462","Type":"ContainerStarted","Data":"f1ce99bf46de76abdc33be766eab5ad4fb04c2e58bf6801c7ae61a8e4e083eaf"} Sep 29 17:21:58 crc kubenswrapper[4667]: I0929 17:21:58.278310 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" event={"ID":"7aba451b-e6ee-41e6-b5f5-ffbcbf9c5462","Type":"ContainerStarted","Data":"1b8d0ddbfad4e13c2bab623dbe836e996edcefd41f5cd2e1af120de5360e9f61"} Sep 29 17:21:58 crc kubenswrapper[4667]: I0929 17:21:58.278671 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:58 crc kubenswrapper[4667]: I0929 17:21:58.285448 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" Sep 29 17:21:58 crc kubenswrapper[4667]: I0929 17:21:58.293040 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66c94b6f9-lnxlc" podStartSLOduration=5.293027166 podStartE2EDuration="5.293027166s" podCreationTimestamp="2025-09-29 17:21:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:21:58.290305785 +0000 UTC m=+746.788152554" watchObservedRunningTime="2025-09-29 17:21:58.293027166 +0000 UTC m=+746.790873924" Sep 29 17:22:02 crc kubenswrapper[4667]: I0929 17:22:02.307197 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" event={"ID":"fbde58f9-2869-40cd-bef8-a28336337936","Type":"ContainerStarted","Data":"560a8730beeffe3cd30c72c9041071d02a798c30bcc926efb04f7610f1fdd99c"} Sep 29 17:22:02 crc kubenswrapper[4667]: I0929 17:22:02.307919 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:22:02 crc kubenswrapper[4667]: I0929 17:22:02.310039 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" Sep 29 17:22:02 crc kubenswrapper[4667]: I0929 17:22:02.331623 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-679b68c759-tgpzd" podStartSLOduration=2.339369048 podStartE2EDuration="14.331607447s" podCreationTimestamp="2025-09-29 17:21:48 +0000 UTC" firstStartedPulling="2025-09-29 17:21:49.300390584 +0000 UTC m=+737.798237352" lastFinishedPulling="2025-09-29 17:22:01.292628982 +0000 UTC m=+749.790475751" observedRunningTime="2025-09-29 17:22:02.327903243 +0000 UTC m=+750.825750013" watchObservedRunningTime="2025-09-29 17:22:02.331607447 +0000 UTC m=+750.829454215" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.028512 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.029641 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.031535 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.031583 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.031893 4667 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-96hr6" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.036884 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.110710 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5f8380cc-fca5-4546-b89f-cb899038ba73\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f8380cc-fca5-4546-b89f-cb899038ba73\") pod \"minio\" (UID: \"d1ac2bba-1994-400e-bff4-f6505b51fea8\") " pod="minio-dev/minio" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.110778 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdbcf\" (UniqueName: \"kubernetes.io/projected/d1ac2bba-1994-400e-bff4-f6505b51fea8-kube-api-access-sdbcf\") pod \"minio\" (UID: \"d1ac2bba-1994-400e-bff4-f6505b51fea8\") " pod="minio-dev/minio" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.211592 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5f8380cc-fca5-4546-b89f-cb899038ba73\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f8380cc-fca5-4546-b89f-cb899038ba73\") pod \"minio\" (UID: \"d1ac2bba-1994-400e-bff4-f6505b51fea8\") " pod="minio-dev/minio" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.211655 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdbcf\" (UniqueName: \"kubernetes.io/projected/d1ac2bba-1994-400e-bff4-f6505b51fea8-kube-api-access-sdbcf\") pod \"minio\" (UID: \"d1ac2bba-1994-400e-bff4-f6505b51fea8\") " pod="minio-dev/minio" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.219306 4667 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.219356 4667 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5f8380cc-fca5-4546-b89f-cb899038ba73\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f8380cc-fca5-4546-b89f-cb899038ba73\") pod \"minio\" (UID: \"d1ac2bba-1994-400e-bff4-f6505b51fea8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/62455b323448611b158259da4ad6636dd4b5fefd859ca7d9d36a621807d1acc5/globalmount\"" pod="minio-dev/minio" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.235634 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdbcf\" (UniqueName: \"kubernetes.io/projected/d1ac2bba-1994-400e-bff4-f6505b51fea8-kube-api-access-sdbcf\") pod \"minio\" (UID: \"d1ac2bba-1994-400e-bff4-f6505b51fea8\") " pod="minio-dev/minio" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.244386 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5f8380cc-fca5-4546-b89f-cb899038ba73\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f8380cc-fca5-4546-b89f-cb899038ba73\") pod \"minio\" (UID: \"d1ac2bba-1994-400e-bff4-f6505b51fea8\") " pod="minio-dev/minio" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.346055 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Sep 29 17:22:06 crc kubenswrapper[4667]: I0929 17:22:06.744646 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Sep 29 17:22:07 crc kubenswrapper[4667]: I0929 17:22:07.335405 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"d1ac2bba-1994-400e-bff4-f6505b51fea8","Type":"ContainerStarted","Data":"4f11261ed55300069c78b40d77738bd4004228988e0a5288781af45294ed7879"} Sep 29 17:22:10 crc kubenswrapper[4667]: I0929 17:22:10.357839 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"d1ac2bba-1994-400e-bff4-f6505b51fea8","Type":"ContainerStarted","Data":"f95edcdc89ac4a7a670d8970e93dc76fa10c16dcb124a92cda5894ccf56481f7"} Sep 29 17:22:10 crc kubenswrapper[4667]: I0929 17:22:10.375582 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.759210966 podStartE2EDuration="7.375544952s" podCreationTimestamp="2025-09-29 17:22:03 +0000 UTC" firstStartedPulling="2025-09-29 17:22:06.752986294 +0000 UTC m=+755.250833063" lastFinishedPulling="2025-09-29 17:22:09.36932028 +0000 UTC m=+757.867167049" observedRunningTime="2025-09-29 17:22:10.370520891 +0000 UTC m=+758.868367671" watchObservedRunningTime="2025-09-29 17:22:10.375544952 +0000 UTC m=+758.873391721" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.308972 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-67c9b4c785-twdch"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.310152 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.312544 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.313796 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.314915 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-bdpnl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.315071 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.315217 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.324160 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-67c9b4c785-twdch"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.433350 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/d60533af-cf3f-4854-acd7-1aa7a11a47e9-logging-loki-distributor-http\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.433550 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/d60533af-cf3f-4854-acd7-1aa7a11a47e9-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.433603 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4sf9\" (UniqueName: \"kubernetes.io/projected/d60533af-cf3f-4854-acd7-1aa7a11a47e9-kube-api-access-t4sf9\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.433643 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d60533af-cf3f-4854-acd7-1aa7a11a47e9-logging-loki-ca-bundle\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.433727 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60533af-cf3f-4854-acd7-1aa7a11a47e9-config\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.478925 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-7454676c57-kv7bp"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.479600 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.481658 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.481855 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.482283 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.491010 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-7454676c57-kv7bp"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.535361 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/d60533af-cf3f-4854-acd7-1aa7a11a47e9-logging-loki-distributor-http\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.535455 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/d60533af-cf3f-4854-acd7-1aa7a11a47e9-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.535485 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-querier-grpc\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.535509 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4sf9\" (UniqueName: \"kubernetes.io/projected/d60533af-cf3f-4854-acd7-1aa7a11a47e9-kube-api-access-t4sf9\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.535537 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036ee610-9a8c-47ee-be84-32a681f82a61-config\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.535562 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-s3\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.535581 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-querier-http\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.535932 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d60533af-cf3f-4854-acd7-1aa7a11a47e9-logging-loki-ca-bundle\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.536053 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60533af-cf3f-4854-acd7-1aa7a11a47e9-config\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.536084 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7klw\" (UniqueName: \"kubernetes.io/projected/036ee610-9a8c-47ee-be84-32a681f82a61-kube-api-access-t7klw\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.536109 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-ca-bundle\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.536792 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d60533af-cf3f-4854-acd7-1aa7a11a47e9-logging-loki-ca-bundle\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.537419 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d60533af-cf3f-4854-acd7-1aa7a11a47e9-config\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.540534 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/d60533af-cf3f-4854-acd7-1aa7a11a47e9-logging-loki-distributor-http\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.553637 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4sf9\" (UniqueName: \"kubernetes.io/projected/d60533af-cf3f-4854-acd7-1aa7a11a47e9-kube-api-access-t4sf9\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.558391 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/d60533af-cf3f-4854-acd7-1aa7a11a47e9-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-67c9b4c785-twdch\" (UID: \"d60533af-cf3f-4854-acd7-1aa7a11a47e9\") " pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.560968 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.561687 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.568072 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.570935 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.576019 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.624195 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638221 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/5d817e8c-29d5-48c6-8596-baddee83a3f3-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638290 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/5d817e8c-29d5-48c6-8596-baddee83a3f3-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638408 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-querier-grpc\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638457 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036ee610-9a8c-47ee-be84-32a681f82a61-config\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638477 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-s3\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638496 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-querier-http\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638533 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d817e8c-29d5-48c6-8596-baddee83a3f3-config\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638559 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d817e8c-29d5-48c6-8596-baddee83a3f3-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638587 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7klw\" (UniqueName: \"kubernetes.io/projected/036ee610-9a8c-47ee-be84-32a681f82a61-kube-api-access-t7klw\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638605 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-ca-bundle\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.638625 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbr2w\" (UniqueName: \"kubernetes.io/projected/5d817e8c-29d5-48c6-8596-baddee83a3f3-kube-api-access-sbr2w\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.639958 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036ee610-9a8c-47ee-be84-32a681f82a61-config\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.642223 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-ca-bundle\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.648947 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-querier-grpc\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.649792 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-s3\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.649983 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/036ee610-9a8c-47ee-be84-32a681f82a61-logging-loki-querier-http\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.659562 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-55cc667c54-fm22d"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.660571 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.663664 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.663733 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.663857 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.663669 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.664086 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-nxt9r" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.664227 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.668744 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7klw\" (UniqueName: \"kubernetes.io/projected/036ee610-9a8c-47ee-be84-32a681f82a61-kube-api-access-t7klw\") pod \"logging-loki-querier-7454676c57-kv7bp\" (UID: \"036ee610-9a8c-47ee-be84-32a681f82a61\") " pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.681626 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-55cc667c54-mv8xl"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.682523 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.687796 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-55cc667c54-mv8xl"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.693419 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-55cc667c54-fm22d"] Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740239 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-tls-secret\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740274 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740306 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-lokistack-gateway\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740421 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-tls-secret\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740466 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-tenants\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740486 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-tenants\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740512 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-rbac\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740531 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-logging-loki-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740578 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/5d817e8c-29d5-48c6-8596-baddee83a3f3-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740764 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740862 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740886 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-rbac\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740917 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/5d817e8c-29d5-48c6-8596-baddee83a3f3-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.740973 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-logging-loki-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.741231 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b42v6\" (UniqueName: \"kubernetes.io/projected/defff6f8-6bee-4b88-b946-32cca2a46fc0-kube-api-access-b42v6\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.741257 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-lokistack-gateway\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.741374 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d817e8c-29d5-48c6-8596-baddee83a3f3-config\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.741405 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d817e8c-29d5-48c6-8596-baddee83a3f3-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.741577 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbr2w\" (UniqueName: \"kubernetes.io/projected/5d817e8c-29d5-48c6-8596-baddee83a3f3-kube-api-access-sbr2w\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.741611 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.741635 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt6rr\" (UniqueName: \"kubernetes.io/projected/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-kube-api-access-kt6rr\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.742403 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5d817e8c-29d5-48c6-8596-baddee83a3f3-config\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.742950 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d817e8c-29d5-48c6-8596-baddee83a3f3-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.743894 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/5d817e8c-29d5-48c6-8596-baddee83a3f3-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.745245 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/5d817e8c-29d5-48c6-8596-baddee83a3f3-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.760326 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbr2w\" (UniqueName: \"kubernetes.io/projected/5d817e8c-29d5-48c6-8596-baddee83a3f3-kube-api-access-sbr2w\") pod \"logging-loki-query-frontend-6b467cdd84-4479c\" (UID: \"5d817e8c-29d5-48c6-8596-baddee83a3f3\") " pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.795345 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.843623 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.843662 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt6rr\" (UniqueName: \"kubernetes.io/projected/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-kube-api-access-kt6rr\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.843692 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.843723 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-tls-secret\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.843744 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-lokistack-gateway\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.843772 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-tls-secret\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.843790 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-tenants\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.843804 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-tenants\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.843822 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-logging-loki-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: E0929 17:22:13.843947 4667 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Sep 29 17:22:13 crc kubenswrapper[4667]: E0929 17:22:13.844029 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-tls-secret podName:defff6f8-6bee-4b88-b946-32cca2a46fc0 nodeName:}" failed. No retries permitted until 2025-09-29 17:22:14.344009258 +0000 UTC m=+762.841856027 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-tls-secret") pod "logging-loki-gateway-55cc667c54-mv8xl" (UID: "defff6f8-6bee-4b88-b946-32cca2a46fc0") : secret "logging-loki-gateway-http" not found Sep 29 17:22:13 crc kubenswrapper[4667]: E0929 17:22:13.844352 4667 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Sep 29 17:22:13 crc kubenswrapper[4667]: E0929 17:22:13.844379 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-tls-secret podName:f50c1454-bcbb-4a46-ab2c-cb717a1083d3 nodeName:}" failed. No retries permitted until 2025-09-29 17:22:14.344371151 +0000 UTC m=+762.842217920 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-tls-secret") pod "logging-loki-gateway-55cc667c54-fm22d" (UID: "f50c1454-bcbb-4a46-ab2c-cb717a1083d3") : secret "logging-loki-gateway-http" not found Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.844605 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.844620 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-rbac\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.844693 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.844746 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.844767 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-rbac\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.844802 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-logging-loki-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.844836 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b42v6\" (UniqueName: \"kubernetes.io/projected/defff6f8-6bee-4b88-b946-32cca2a46fc0-kube-api-access-b42v6\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.844875 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-lokistack-gateway\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.844797 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-logging-loki-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.845643 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.845671 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-lokistack-gateway\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.845978 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-logging-loki-ca-bundle\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.846273 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-rbac\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.846477 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-rbac\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.847109 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-tenants\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.847164 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-tenants\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.849435 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/defff6f8-6bee-4b88-b946-32cca2a46fc0-lokistack-gateway\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.853413 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.854196 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.859266 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt6rr\" (UniqueName: \"kubernetes.io/projected/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-kube-api-access-kt6rr\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.861645 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b42v6\" (UniqueName: \"kubernetes.io/projected/defff6f8-6bee-4b88-b946-32cca2a46fc0-kube-api-access-b42v6\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:13 crc kubenswrapper[4667]: I0929 17:22:13.900822 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.070881 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-67c9b4c785-twdch"] Sep 29 17:22:14 crc kubenswrapper[4667]: W0929 17:22:14.074101 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd60533af_cf3f_4854_acd7_1aa7a11a47e9.slice/crio-310f48b09ceac3ae2366e77f411f2da5562e040fdaddb528a8597d4b26a7dfc6 WatchSource:0}: Error finding container 310f48b09ceac3ae2366e77f411f2da5562e040fdaddb528a8597d4b26a7dfc6: Status 404 returned error can't find the container with id 310f48b09ceac3ae2366e77f411f2da5562e040fdaddb528a8597d4b26a7dfc6 Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.176507 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-7454676c57-kv7bp"] Sep 29 17:22:14 crc kubenswrapper[4667]: W0929 17:22:14.181510 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod036ee610_9a8c_47ee_be84_32a681f82a61.slice/crio-36fdb6a81ade87e3bf7477865b65bf32065c90e22a9bfeff0eb1030d45957e2a WatchSource:0}: Error finding container 36fdb6a81ade87e3bf7477865b65bf32065c90e22a9bfeff0eb1030d45957e2a: Status 404 returned error can't find the container with id 36fdb6a81ade87e3bf7477865b65bf32065c90e22a9bfeff0eb1030d45957e2a Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.314317 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c"] Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.355337 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-tls-secret\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.355411 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-tls-secret\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.360382 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/defff6f8-6bee-4b88-b946-32cca2a46fc0-tls-secret\") pod \"logging-loki-gateway-55cc667c54-mv8xl\" (UID: \"defff6f8-6bee-4b88-b946-32cca2a46fc0\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.360388 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/f50c1454-bcbb-4a46-ab2c-cb717a1083d3-tls-secret\") pod \"logging-loki-gateway-55cc667c54-fm22d\" (UID: \"f50c1454-bcbb-4a46-ab2c-cb717a1083d3\") " pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.383466 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" event={"ID":"5d817e8c-29d5-48c6-8596-baddee83a3f3","Type":"ContainerStarted","Data":"b31abb7c9e41824645097fb4a0a9859931c57fb94596bc82b1d354941f9efdb2"} Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.384532 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" event={"ID":"d60533af-cf3f-4854-acd7-1aa7a11a47e9","Type":"ContainerStarted","Data":"310f48b09ceac3ae2366e77f411f2da5562e040fdaddb528a8597d4b26a7dfc6"} Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.385620 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" event={"ID":"036ee610-9a8c-47ee-be84-32a681f82a61","Type":"ContainerStarted","Data":"36fdb6a81ade87e3bf7477865b65bf32065c90e22a9bfeff0eb1030d45957e2a"} Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.471146 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.472069 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.473898 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.474210 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.481310 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.543772 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.544867 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.546352 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.546500 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.553746 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.567973 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-17af033c-6874-4eed-a421-7690fd97468b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-17af033c-6874-4eed-a421-7690fd97468b\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.568042 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qvcj\" (UniqueName: \"kubernetes.io/projected/658afe18-7fa3-4e33-bf5c-cfab27925cb4-kube-api-access-7qvcj\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.568067 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.568164 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.568200 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4b104cff-9d0c-4457-9da9-8fed86ce828b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b104cff-9d0c-4457-9da9-8fed86ce828b\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.568247 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.568344 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/658afe18-7fa3-4e33-bf5c-cfab27925cb4-config\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.568377 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.598102 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.604775 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.620158 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.621225 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.624823 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.625114 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.628720 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.670749 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qvcj\" (UniqueName: \"kubernetes.io/projected/658afe18-7fa3-4e33-bf5c-cfab27925cb4-kube-api-access-7qvcj\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.670966 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.670999 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-42936075-a85e-47b1-b6d9-f739b45d7598\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42936075-a85e-47b1-b6d9-f739b45d7598\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671024 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671067 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671084 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/647a9667-397f-4cb5-91bf-da76ac52ccad-config\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671105 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671127 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671149 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4b104cff-9d0c-4457-9da9-8fed86ce828b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b104cff-9d0c-4457-9da9-8fed86ce828b\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671174 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671201 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671270 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/658afe18-7fa3-4e33-bf5c-cfab27925cb4-config\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671290 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671327 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rblns\" (UniqueName: \"kubernetes.io/projected/647a9667-397f-4cb5-91bf-da76ac52ccad-kube-api-access-rblns\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.671352 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-17af033c-6874-4eed-a421-7690fd97468b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-17af033c-6874-4eed-a421-7690fd97468b\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.673264 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/658afe18-7fa3-4e33-bf5c-cfab27925cb4-config\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.673348 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.675954 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.676544 4667 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.676554 4667 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.676574 4667 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4b104cff-9d0c-4457-9da9-8fed86ce828b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b104cff-9d0c-4457-9da9-8fed86ce828b\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/30f7861dac16e4109c4f1a15f527801a30a25b750dedde726aff3743ae34c373/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.676577 4667 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-17af033c-6874-4eed-a421-7690fd97468b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-17af033c-6874-4eed-a421-7690fd97468b\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c6111c97c903619ed6b1115064dfeead1fe96f2d034d6d283093b664ad2b3bab/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.684809 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.686709 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/658afe18-7fa3-4e33-bf5c-cfab27925cb4-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.688520 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qvcj\" (UniqueName: \"kubernetes.io/projected/658afe18-7fa3-4e33-bf5c-cfab27925cb4-kube-api-access-7qvcj\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.697295 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4b104cff-9d0c-4457-9da9-8fed86ce828b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b104cff-9d0c-4457-9da9-8fed86ce828b\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.704296 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-17af033c-6874-4eed-a421-7690fd97468b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-17af033c-6874-4eed-a421-7690fd97468b\") pod \"logging-loki-ingester-0\" (UID: \"658afe18-7fa3-4e33-bf5c-cfab27925cb4\") " pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.773655 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-config\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.773721 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.773759 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.773791 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74v4x\" (UniqueName: \"kubernetes.io/projected/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-kube-api-access-74v4x\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.773822 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rblns\" (UniqueName: \"kubernetes.io/projected/647a9667-397f-4cb5-91bf-da76ac52ccad-kube-api-access-rblns\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.773872 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-42936075-a85e-47b1-b6d9-f739b45d7598\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42936075-a85e-47b1-b6d9-f739b45d7598\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.773910 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.773939 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.773989 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.774019 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3ce1c2aa-08da-4b78-b11b-0e42ce4ab3cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3ce1c2aa-08da-4b78-b11b-0e42ce4ab3cc\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.774049 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.774082 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/647a9667-397f-4cb5-91bf-da76ac52ccad-config\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.774106 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.774135 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.777635 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/647a9667-397f-4cb5-91bf-da76ac52ccad-config\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.778754 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.778945 4667 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.778977 4667 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-42936075-a85e-47b1-b6d9-f739b45d7598\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42936075-a85e-47b1-b6d9-f739b45d7598\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8de2564ca3b1dc2bfe79a42dde0074d60838babe41d231529c50fe1f5eccb3a8/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.780177 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.780169 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.781360 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/647a9667-397f-4cb5-91bf-da76ac52ccad-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.790146 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rblns\" (UniqueName: \"kubernetes.io/projected/647a9667-397f-4cb5-91bf-da76ac52ccad-kube-api-access-rblns\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.790878 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.799978 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-42936075-a85e-47b1-b6d9-f739b45d7598\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-42936075-a85e-47b1-b6d9-f739b45d7598\") pod \"logging-loki-compactor-0\" (UID: \"647a9667-397f-4cb5-91bf-da76ac52ccad\") " pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.879075 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74v4x\" (UniqueName: \"kubernetes.io/projected/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-kube-api-access-74v4x\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.879249 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.879291 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3ce1c2aa-08da-4b78-b11b-0e42ce4ab3cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3ce1c2aa-08da-4b78-b11b-0e42ce4ab3cc\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.879372 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.879407 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-config\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.879446 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.879472 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.880325 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.880784 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-config\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.884053 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.884390 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.885104 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.894070 4667 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.894103 4667 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3ce1c2aa-08da-4b78-b11b-0e42ce4ab3cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3ce1c2aa-08da-4b78-b11b-0e42ce4ab3cc\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e4ed443dfb96fb4517db6f9d736c0cb16871dc5ddc9837f9e02d15d24adc42f6/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.896299 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.896621 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74v4x\" (UniqueName: \"kubernetes.io/projected/980ab4d7-2edd-46f3-9f6c-b6138f1c3350-kube-api-access-74v4x\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.920597 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3ce1c2aa-08da-4b78-b11b-0e42ce4ab3cc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3ce1c2aa-08da-4b78-b11b-0e42ce4ab3cc\") pod \"logging-loki-index-gateway-0\" (UID: \"980ab4d7-2edd-46f3-9f6c-b6138f1c3350\") " pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.957404 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:14 crc kubenswrapper[4667]: I0929 17:22:14.985628 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-55cc667c54-fm22d"] Sep 29 17:22:14 crc kubenswrapper[4667]: W0929 17:22:14.993200 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf50c1454_bcbb_4a46_ab2c_cb717a1083d3.slice/crio-4c5218d91ee04b3ff15fd694ce5b1fa7913c5306cbd9362c0625bda81cd01ae8 WatchSource:0}: Error finding container 4c5218d91ee04b3ff15fd694ce5b1fa7913c5306cbd9362c0625bda81cd01ae8: Status 404 returned error can't find the container with id 4c5218d91ee04b3ff15fd694ce5b1fa7913c5306cbd9362c0625bda81cd01ae8 Sep 29 17:22:15 crc kubenswrapper[4667]: I0929 17:22:15.067089 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-55cc667c54-mv8xl"] Sep 29 17:22:15 crc kubenswrapper[4667]: I0929 17:22:15.218168 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Sep 29 17:22:15 crc kubenswrapper[4667]: I0929 17:22:15.285763 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Sep 29 17:22:15 crc kubenswrapper[4667]: W0929 17:22:15.292914 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod647a9667_397f_4cb5_91bf_da76ac52ccad.slice/crio-a80e94369ce3821dd783a7123a87457d87091e17f0c3687f3229b52fa3f25c57 WatchSource:0}: Error finding container a80e94369ce3821dd783a7123a87457d87091e17f0c3687f3229b52fa3f25c57: Status 404 returned error can't find the container with id a80e94369ce3821dd783a7123a87457d87091e17f0c3687f3229b52fa3f25c57 Sep 29 17:22:15 crc kubenswrapper[4667]: I0929 17:22:15.393668 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"658afe18-7fa3-4e33-bf5c-cfab27925cb4","Type":"ContainerStarted","Data":"f3a33e9022a21d7522e0ffdb77619de8c2ea2fc9c0ffbf1e33e39193be0b4d50"} Sep 29 17:22:15 crc kubenswrapper[4667]: I0929 17:22:15.394824 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"647a9667-397f-4cb5-91bf-da76ac52ccad","Type":"ContainerStarted","Data":"a80e94369ce3821dd783a7123a87457d87091e17f0c3687f3229b52fa3f25c57"} Sep 29 17:22:15 crc kubenswrapper[4667]: I0929 17:22:15.396055 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" event={"ID":"defff6f8-6bee-4b88-b946-32cca2a46fc0","Type":"ContainerStarted","Data":"f255df16b61ca54ade7176209c5930164376334d934d569fc9924d4c67bd5b69"} Sep 29 17:22:15 crc kubenswrapper[4667]: I0929 17:22:15.396927 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" event={"ID":"f50c1454-bcbb-4a46-ab2c-cb717a1083d3","Type":"ContainerStarted","Data":"4c5218d91ee04b3ff15fd694ce5b1fa7913c5306cbd9362c0625bda81cd01ae8"} Sep 29 17:22:15 crc kubenswrapper[4667]: I0929 17:22:15.432306 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Sep 29 17:22:16 crc kubenswrapper[4667]: I0929 17:22:16.405213 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"980ab4d7-2edd-46f3-9f6c-b6138f1c3350","Type":"ContainerStarted","Data":"6a020a6cf1ac2c6eb7ccfb2909218127bd2bb6ba875e2c1bff160e77c7b47468"} Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.421890 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"658afe18-7fa3-4e33-bf5c-cfab27925cb4","Type":"ContainerStarted","Data":"28284da660ece6af379288a45e305e94f533a315f89a9970e22a061aa23715ed"} Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.423132 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.425056 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" event={"ID":"5d817e8c-29d5-48c6-8596-baddee83a3f3","Type":"ContainerStarted","Data":"90584c92c2b20d9165823b8d41352d3761940daa8c11b817a85d0324887285d6"} Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.425289 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.431693 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" event={"ID":"d60533af-cf3f-4854-acd7-1aa7a11a47e9","Type":"ContainerStarted","Data":"b3be65bb8bcf8c725d5d3cf31a20c6c10f1127f0143a59364bf8765c9f5d2273"} Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.431765 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.437756 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"647a9667-397f-4cb5-91bf-da76ac52ccad","Type":"ContainerStarted","Data":"b0a8d2021f9d6274352f226244ba5ae3df2a4a32b48bffd8f77dc1b160ccef82"} Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.438025 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.446620 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" event={"ID":"036ee610-9a8c-47ee-be84-32a681f82a61","Type":"ContainerStarted","Data":"d59c7c630dfb19fae56edf8c1eca54c5d2a0e02f3ee96894dc33bf08a16a8f60"} Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.446993 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.449454 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"980ab4d7-2edd-46f3-9f6c-b6138f1c3350","Type":"ContainerStarted","Data":"7d4b0df497645f2e85cf65aab32aabc2e58cfd2dfe350f9b97b707447a3de0cb"} Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.449991 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.451123 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=3.4658783890000002 podStartE2EDuration="5.451103886s" podCreationTimestamp="2025-09-29 17:22:13 +0000 UTC" firstStartedPulling="2025-09-29 17:22:15.257349816 +0000 UTC m=+763.755196586" lastFinishedPulling="2025-09-29 17:22:17.242575324 +0000 UTC m=+765.740422083" observedRunningTime="2025-09-29 17:22:18.439736779 +0000 UTC m=+766.937583547" watchObservedRunningTime="2025-09-29 17:22:18.451103886 +0000 UTC m=+766.948950655" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.473198 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.555109702 podStartE2EDuration="5.473180991s" podCreationTimestamp="2025-09-29 17:22:13 +0000 UTC" firstStartedPulling="2025-09-29 17:22:15.29510121 +0000 UTC m=+763.792947978" lastFinishedPulling="2025-09-29 17:22:17.213172498 +0000 UTC m=+765.711019267" observedRunningTime="2025-09-29 17:22:18.459068499 +0000 UTC m=+766.956915268" watchObservedRunningTime="2025-09-29 17:22:18.473180991 +0000 UTC m=+766.971027761" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.484740 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" podStartSLOduration=2.346191827 podStartE2EDuration="5.484725584s" podCreationTimestamp="2025-09-29 17:22:13 +0000 UTC" firstStartedPulling="2025-09-29 17:22:14.08146523 +0000 UTC m=+762.579311999" lastFinishedPulling="2025-09-29 17:22:17.219998987 +0000 UTC m=+765.717845756" observedRunningTime="2025-09-29 17:22:18.476167742 +0000 UTC m=+766.974014500" watchObservedRunningTime="2025-09-29 17:22:18.484725584 +0000 UTC m=+766.982572352" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.490291 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" podStartSLOduration=2.589295764 podStartE2EDuration="5.490276086s" podCreationTimestamp="2025-09-29 17:22:13 +0000 UTC" firstStartedPulling="2025-09-29 17:22:14.319916678 +0000 UTC m=+762.817763447" lastFinishedPulling="2025-09-29 17:22:17.220897 +0000 UTC m=+765.718743769" observedRunningTime="2025-09-29 17:22:18.486374532 +0000 UTC m=+766.984221291" watchObservedRunningTime="2025-09-29 17:22:18.490276086 +0000 UTC m=+766.988122855" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.501445 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" podStartSLOduration=2.441962016 podStartE2EDuration="5.501431515s" podCreationTimestamp="2025-09-29 17:22:13 +0000 UTC" firstStartedPulling="2025-09-29 17:22:14.183422913 +0000 UTC m=+762.681269682" lastFinishedPulling="2025-09-29 17:22:17.242892412 +0000 UTC m=+765.740739181" observedRunningTime="2025-09-29 17:22:18.498152294 +0000 UTC m=+766.995999062" watchObservedRunningTime="2025-09-29 17:22:18.501431515 +0000 UTC m=+766.999278283" Sep 29 17:22:18 crc kubenswrapper[4667]: I0929 17:22:18.524177 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.652187606 podStartE2EDuration="5.524160018s" podCreationTimestamp="2025-09-29 17:22:13 +0000 UTC" firstStartedPulling="2025-09-29 17:22:15.420720575 +0000 UTC m=+763.918567344" lastFinishedPulling="2025-09-29 17:22:17.292692987 +0000 UTC m=+765.790539756" observedRunningTime="2025-09-29 17:22:18.524008293 +0000 UTC m=+767.021855062" watchObservedRunningTime="2025-09-29 17:22:18.524160018 +0000 UTC m=+767.022006787" Sep 29 17:22:19 crc kubenswrapper[4667]: I0929 17:22:19.455648 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" event={"ID":"f50c1454-bcbb-4a46-ab2c-cb717a1083d3","Type":"ContainerStarted","Data":"c0ada41455f79c791e924bfb036f62e3611184060764eafffbba96a67501b242"} Sep 29 17:22:19 crc kubenswrapper[4667]: I0929 17:22:19.457348 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" event={"ID":"defff6f8-6bee-4b88-b946-32cca2a46fc0","Type":"ContainerStarted","Data":"c75873f27c87e40c05ed748f0aaf4336729a6d107921b21adee9ca54a4d03bf8"} Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.469912 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" event={"ID":"f50c1454-bcbb-4a46-ab2c-cb717a1083d3","Type":"ContainerStarted","Data":"2e1fae60d2949670b68d99baa50829ef2be246f1df6debe0e804db674f1258ef"} Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.470319 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.470333 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.471976 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" event={"ID":"defff6f8-6bee-4b88-b946-32cca2a46fc0","Type":"ContainerStarted","Data":"bd86e17d25d140b95a21730bdc8559414b34407b93d2f2a24c5798c4d69d0d97"} Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.472242 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.472266 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.478725 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.479046 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.480719 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.483455 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.489467 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-55cc667c54-fm22d" podStartSLOduration=3.082085838 podStartE2EDuration="8.48945709s" podCreationTimestamp="2025-09-29 17:22:13 +0000 UTC" firstStartedPulling="2025-09-29 17:22:14.99622578 +0000 UTC m=+763.494072549" lastFinishedPulling="2025-09-29 17:22:20.403597032 +0000 UTC m=+768.901443801" observedRunningTime="2025-09-29 17:22:21.488032274 +0000 UTC m=+769.985879043" watchObservedRunningTime="2025-09-29 17:22:21.48945709 +0000 UTC m=+769.987303859" Sep 29 17:22:21 crc kubenswrapper[4667]: I0929 17:22:21.504658 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-55cc667c54-mv8xl" podStartSLOduration=3.178080028 podStartE2EDuration="8.504643997s" podCreationTimestamp="2025-09-29 17:22:13 +0000 UTC" firstStartedPulling="2025-09-29 17:22:15.072445356 +0000 UTC m=+763.570292125" lastFinishedPulling="2025-09-29 17:22:20.399009324 +0000 UTC m=+768.896856094" observedRunningTime="2025-09-29 17:22:21.502951216 +0000 UTC m=+770.000797985" watchObservedRunningTime="2025-09-29 17:22:21.504643997 +0000 UTC m=+770.002490767" Sep 29 17:22:26 crc kubenswrapper[4667]: I0929 17:22:26.897295 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:22:26 crc kubenswrapper[4667]: I0929 17:22:26.897606 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.081586 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6vbqf"] Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.084213 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.092493 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vbqf"] Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.181075 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6z5w\" (UniqueName: \"kubernetes.io/projected/3d3eaf62-b0e0-439e-b085-86af16c929b4-kube-api-access-r6z5w\") pod \"redhat-marketplace-6vbqf\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.181269 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-utilities\") pod \"redhat-marketplace-6vbqf\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.181521 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-catalog-content\") pod \"redhat-marketplace-6vbqf\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.283079 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-utilities\") pod \"redhat-marketplace-6vbqf\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.283204 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-catalog-content\") pod \"redhat-marketplace-6vbqf\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.283514 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6z5w\" (UniqueName: \"kubernetes.io/projected/3d3eaf62-b0e0-439e-b085-86af16c929b4-kube-api-access-r6z5w\") pod \"redhat-marketplace-6vbqf\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.283610 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-catalog-content\") pod \"redhat-marketplace-6vbqf\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.283889 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-utilities\") pod \"redhat-marketplace-6vbqf\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.316744 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6z5w\" (UniqueName: \"kubernetes.io/projected/3d3eaf62-b0e0-439e-b085-86af16c929b4-kube-api-access-r6z5w\") pod \"redhat-marketplace-6vbqf\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.403784 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:32 crc kubenswrapper[4667]: I0929 17:22:32.750646 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vbqf"] Sep 29 17:22:32 crc kubenswrapper[4667]: W0929 17:22:32.753198 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d3eaf62_b0e0_439e_b085_86af16c929b4.slice/crio-7aad58da6f46c02b7ee7457ec295ea00ba5072c906a8a5094071604859b7ef3a WatchSource:0}: Error finding container 7aad58da6f46c02b7ee7457ec295ea00ba5072c906a8a5094071604859b7ef3a: Status 404 returned error can't find the container with id 7aad58da6f46c02b7ee7457ec295ea00ba5072c906a8a5094071604859b7ef3a Sep 29 17:22:33 crc kubenswrapper[4667]: I0929 17:22:33.538531 4667 generic.go:334] "Generic (PLEG): container finished" podID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerID="4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf" exitCode=0 Sep 29 17:22:33 crc kubenswrapper[4667]: I0929 17:22:33.538583 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vbqf" event={"ID":"3d3eaf62-b0e0-439e-b085-86af16c929b4","Type":"ContainerDied","Data":"4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf"} Sep 29 17:22:33 crc kubenswrapper[4667]: I0929 17:22:33.538792 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vbqf" event={"ID":"3d3eaf62-b0e0-439e-b085-86af16c929b4","Type":"ContainerStarted","Data":"7aad58da6f46c02b7ee7457ec295ea00ba5072c906a8a5094071604859b7ef3a"} Sep 29 17:22:33 crc kubenswrapper[4667]: I0929 17:22:33.629644 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-67c9b4c785-twdch" Sep 29 17:22:33 crc kubenswrapper[4667]: I0929 17:22:33.800672 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-7454676c57-kv7bp" Sep 29 17:22:33 crc kubenswrapper[4667]: I0929 17:22:33.906096 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-6b467cdd84-4479c" Sep 29 17:22:34 crc kubenswrapper[4667]: I0929 17:22:34.545068 4667 generic.go:334] "Generic (PLEG): container finished" podID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerID="81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707" exitCode=0 Sep 29 17:22:34 crc kubenswrapper[4667]: I0929 17:22:34.545104 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vbqf" event={"ID":"3d3eaf62-b0e0-439e-b085-86af16c929b4","Type":"ContainerDied","Data":"81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707"} Sep 29 17:22:34 crc kubenswrapper[4667]: I0929 17:22:34.795757 4667 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Sep 29 17:22:34 crc kubenswrapper[4667]: I0929 17:22:34.795795 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="658afe18-7fa3-4e33-bf5c-cfab27925cb4" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 29 17:22:34 crc kubenswrapper[4667]: I0929 17:22:34.889586 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Sep 29 17:22:34 crc kubenswrapper[4667]: I0929 17:22:34.965357 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Sep 29 17:22:35 crc kubenswrapper[4667]: I0929 17:22:35.552131 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vbqf" event={"ID":"3d3eaf62-b0e0-439e-b085-86af16c929b4","Type":"ContainerStarted","Data":"4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51"} Sep 29 17:22:35 crc kubenswrapper[4667]: I0929 17:22:35.565903 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6vbqf" podStartSLOduration=1.895422946 podStartE2EDuration="3.565890547s" podCreationTimestamp="2025-09-29 17:22:32 +0000 UTC" firstStartedPulling="2025-09-29 17:22:33.539886669 +0000 UTC m=+782.037733439" lastFinishedPulling="2025-09-29 17:22:35.210354271 +0000 UTC m=+783.708201040" observedRunningTime="2025-09-29 17:22:35.563153747 +0000 UTC m=+784.061000516" watchObservedRunningTime="2025-09-29 17:22:35.565890547 +0000 UTC m=+784.063737326" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.047304 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nll2m"] Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.048632 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.054005 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nll2m"] Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.163578 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-catalog-content\") pod \"redhat-operators-nll2m\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.163879 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqbxf\" (UniqueName: \"kubernetes.io/projected/a8713a8f-02d7-4b54-aeb7-b9d09d017543-kube-api-access-sqbxf\") pod \"redhat-operators-nll2m\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.164057 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-utilities\") pod \"redhat-operators-nll2m\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.265767 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-catalog-content\") pod \"redhat-operators-nll2m\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.266046 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqbxf\" (UniqueName: \"kubernetes.io/projected/a8713a8f-02d7-4b54-aeb7-b9d09d017543-kube-api-access-sqbxf\") pod \"redhat-operators-nll2m\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.266180 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-catalog-content\") pod \"redhat-operators-nll2m\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.266272 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-utilities\") pod \"redhat-operators-nll2m\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.266467 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-utilities\") pod \"redhat-operators-nll2m\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.280761 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqbxf\" (UniqueName: \"kubernetes.io/projected/a8713a8f-02d7-4b54-aeb7-b9d09d017543-kube-api-access-sqbxf\") pod \"redhat-operators-nll2m\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.362480 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:39 crc kubenswrapper[4667]: I0929 17:22:39.723342 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nll2m"] Sep 29 17:22:39 crc kubenswrapper[4667]: W0929 17:22:39.724600 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8713a8f_02d7_4b54_aeb7_b9d09d017543.slice/crio-331a6dfa36396b179b8127f524c2a0c2bb3701bb8b0906e938a28391ab17b244 WatchSource:0}: Error finding container 331a6dfa36396b179b8127f524c2a0c2bb3701bb8b0906e938a28391ab17b244: Status 404 returned error can't find the container with id 331a6dfa36396b179b8127f524c2a0c2bb3701bb8b0906e938a28391ab17b244 Sep 29 17:22:40 crc kubenswrapper[4667]: I0929 17:22:40.575789 4667 generic.go:334] "Generic (PLEG): container finished" podID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerID="8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8" exitCode=0 Sep 29 17:22:40 crc kubenswrapper[4667]: I0929 17:22:40.575955 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nll2m" event={"ID":"a8713a8f-02d7-4b54-aeb7-b9d09d017543","Type":"ContainerDied","Data":"8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8"} Sep 29 17:22:40 crc kubenswrapper[4667]: I0929 17:22:40.575998 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nll2m" event={"ID":"a8713a8f-02d7-4b54-aeb7-b9d09d017543","Type":"ContainerStarted","Data":"331a6dfa36396b179b8127f524c2a0c2bb3701bb8b0906e938a28391ab17b244"} Sep 29 17:22:41 crc kubenswrapper[4667]: I0929 17:22:41.581578 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nll2m" event={"ID":"a8713a8f-02d7-4b54-aeb7-b9d09d017543","Type":"ContainerStarted","Data":"16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc"} Sep 29 17:22:42 crc kubenswrapper[4667]: I0929 17:22:42.405121 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:42 crc kubenswrapper[4667]: I0929 17:22:42.405166 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:42 crc kubenswrapper[4667]: I0929 17:22:42.435996 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:42 crc kubenswrapper[4667]: I0929 17:22:42.587880 4667 generic.go:334] "Generic (PLEG): container finished" podID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerID="16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc" exitCode=0 Sep 29 17:22:42 crc kubenswrapper[4667]: I0929 17:22:42.587978 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nll2m" event={"ID":"a8713a8f-02d7-4b54-aeb7-b9d09d017543","Type":"ContainerDied","Data":"16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc"} Sep 29 17:22:42 crc kubenswrapper[4667]: I0929 17:22:42.622802 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:43 crc kubenswrapper[4667]: I0929 17:22:43.595192 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nll2m" event={"ID":"a8713a8f-02d7-4b54-aeb7-b9d09d017543","Type":"ContainerStarted","Data":"4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7"} Sep 29 17:22:43 crc kubenswrapper[4667]: I0929 17:22:43.607794 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nll2m" podStartSLOduration=1.9892769430000001 podStartE2EDuration="4.607779165s" podCreationTimestamp="2025-09-29 17:22:39 +0000 UTC" firstStartedPulling="2025-09-29 17:22:40.577112194 +0000 UTC m=+789.074958962" lastFinishedPulling="2025-09-29 17:22:43.195614415 +0000 UTC m=+791.693461184" observedRunningTime="2025-09-29 17:22:43.606136047 +0000 UTC m=+792.103982816" watchObservedRunningTime="2025-09-29 17:22:43.607779165 +0000 UTC m=+792.105625934" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.665310 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kfwz2"] Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.666659 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.673103 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfwz2"] Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.798876 4667 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.798919 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="658afe18-7fa3-4e33-bf5c-cfab27925cb4" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.837510 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r22h\" (UniqueName: \"kubernetes.io/projected/b017e67d-043f-4fc3-b342-6480df334c2b-kube-api-access-8r22h\") pod \"certified-operators-kfwz2\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.837550 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-catalog-content\") pod \"certified-operators-kfwz2\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.837588 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-utilities\") pod \"certified-operators-kfwz2\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.865432 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vbqf"] Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.865661 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6vbqf" podUID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerName="registry-server" containerID="cri-o://4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51" gracePeriod=2 Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.939179 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r22h\" (UniqueName: \"kubernetes.io/projected/b017e67d-043f-4fc3-b342-6480df334c2b-kube-api-access-8r22h\") pod \"certified-operators-kfwz2\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.939229 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-catalog-content\") pod \"certified-operators-kfwz2\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.939279 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-utilities\") pod \"certified-operators-kfwz2\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.939975 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-utilities\") pod \"certified-operators-kfwz2\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.940202 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-catalog-content\") pod \"certified-operators-kfwz2\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.964469 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r22h\" (UniqueName: \"kubernetes.io/projected/b017e67d-043f-4fc3-b342-6480df334c2b-kube-api-access-8r22h\") pod \"certified-operators-kfwz2\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:44 crc kubenswrapper[4667]: I0929 17:22:44.979446 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.355719 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.447972 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6z5w\" (UniqueName: \"kubernetes.io/projected/3d3eaf62-b0e0-439e-b085-86af16c929b4-kube-api-access-r6z5w\") pod \"3d3eaf62-b0e0-439e-b085-86af16c929b4\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.448035 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-utilities\") pod \"3d3eaf62-b0e0-439e-b085-86af16c929b4\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.448083 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-catalog-content\") pod \"3d3eaf62-b0e0-439e-b085-86af16c929b4\" (UID: \"3d3eaf62-b0e0-439e-b085-86af16c929b4\") " Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.448785 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-utilities" (OuterVolumeSpecName: "utilities") pod "3d3eaf62-b0e0-439e-b085-86af16c929b4" (UID: "3d3eaf62-b0e0-439e-b085-86af16c929b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.451402 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d3eaf62-b0e0-439e-b085-86af16c929b4-kube-api-access-r6z5w" (OuterVolumeSpecName: "kube-api-access-r6z5w") pod "3d3eaf62-b0e0-439e-b085-86af16c929b4" (UID: "3d3eaf62-b0e0-439e-b085-86af16c929b4"). InnerVolumeSpecName "kube-api-access-r6z5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.457655 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d3eaf62-b0e0-439e-b085-86af16c929b4" (UID: "3d3eaf62-b0e0-439e-b085-86af16c929b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.532837 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfwz2"] Sep 29 17:22:45 crc kubenswrapper[4667]: W0929 17:22:45.536043 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb017e67d_043f_4fc3_b342_6480df334c2b.slice/crio-36481a59d3509185688dc305c9b793d7702e0ff6df7e29fa58671810b39ca288 WatchSource:0}: Error finding container 36481a59d3509185688dc305c9b793d7702e0ff6df7e29fa58671810b39ca288: Status 404 returned error can't find the container with id 36481a59d3509185688dc305c9b793d7702e0ff6df7e29fa58671810b39ca288 Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.550129 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.550154 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d3eaf62-b0e0-439e-b085-86af16c929b4-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.550168 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6z5w\" (UniqueName: \"kubernetes.io/projected/3d3eaf62-b0e0-439e-b085-86af16c929b4-kube-api-access-r6z5w\") on node \"crc\" DevicePath \"\"" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.606590 4667 generic.go:334] "Generic (PLEG): container finished" podID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerID="4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51" exitCode=0 Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.606635 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6vbqf" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.606651 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vbqf" event={"ID":"3d3eaf62-b0e0-439e-b085-86af16c929b4","Type":"ContainerDied","Data":"4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51"} Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.606986 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6vbqf" event={"ID":"3d3eaf62-b0e0-439e-b085-86af16c929b4","Type":"ContainerDied","Data":"7aad58da6f46c02b7ee7457ec295ea00ba5072c906a8a5094071604859b7ef3a"} Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.607027 4667 scope.go:117] "RemoveContainer" containerID="4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.607699 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwz2" event={"ID":"b017e67d-043f-4fc3-b342-6480df334c2b","Type":"ContainerStarted","Data":"36481a59d3509185688dc305c9b793d7702e0ff6df7e29fa58671810b39ca288"} Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.619118 4667 scope.go:117] "RemoveContainer" containerID="81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.631646 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vbqf"] Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.632073 4667 scope.go:117] "RemoveContainer" containerID="4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.635214 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6vbqf"] Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.660290 4667 scope.go:117] "RemoveContainer" containerID="4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51" Sep 29 17:22:45 crc kubenswrapper[4667]: E0929 17:22:45.660589 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51\": container with ID starting with 4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51 not found: ID does not exist" containerID="4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.660623 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51"} err="failed to get container status \"4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51\": rpc error: code = NotFound desc = could not find container \"4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51\": container with ID starting with 4c0b059f0fb5c51268e2afcfc041504183f015023aa392a7af10aeaae58e4d51 not found: ID does not exist" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.660643 4667 scope.go:117] "RemoveContainer" containerID="81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707" Sep 29 17:22:45 crc kubenswrapper[4667]: E0929 17:22:45.660921 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707\": container with ID starting with 81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707 not found: ID does not exist" containerID="81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.660947 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707"} err="failed to get container status \"81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707\": rpc error: code = NotFound desc = could not find container \"81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707\": container with ID starting with 81b83c218b45fd22e1afd17b90a0db8f3043339f683f864c93b05bce562af707 not found: ID does not exist" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.660966 4667 scope.go:117] "RemoveContainer" containerID="4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf" Sep 29 17:22:45 crc kubenswrapper[4667]: E0929 17:22:45.661240 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf\": container with ID starting with 4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf not found: ID does not exist" containerID="4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.661274 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf"} err="failed to get container status \"4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf\": rpc error: code = NotFound desc = could not find container \"4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf\": container with ID starting with 4ff5ce1ad0803fffcefbd119b34a373b52795d0d95ef032397064a00be8b71bf not found: ID does not exist" Sep 29 17:22:45 crc kubenswrapper[4667]: I0929 17:22:45.822412 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d3eaf62-b0e0-439e-b085-86af16c929b4" path="/var/lib/kubelet/pods/3d3eaf62-b0e0-439e-b085-86af16c929b4/volumes" Sep 29 17:22:46 crc kubenswrapper[4667]: I0929 17:22:46.615476 4667 generic.go:334] "Generic (PLEG): container finished" podID="b017e67d-043f-4fc3-b342-6480df334c2b" containerID="6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a" exitCode=0 Sep 29 17:22:46 crc kubenswrapper[4667]: I0929 17:22:46.615529 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwz2" event={"ID":"b017e67d-043f-4fc3-b342-6480df334c2b","Type":"ContainerDied","Data":"6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a"} Sep 29 17:22:47 crc kubenswrapper[4667]: I0929 17:22:47.622082 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwz2" event={"ID":"b017e67d-043f-4fc3-b342-6480df334c2b","Type":"ContainerStarted","Data":"d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983"} Sep 29 17:22:48 crc kubenswrapper[4667]: I0929 17:22:48.627764 4667 generic.go:334] "Generic (PLEG): container finished" podID="b017e67d-043f-4fc3-b342-6480df334c2b" containerID="d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983" exitCode=0 Sep 29 17:22:48 crc kubenswrapper[4667]: I0929 17:22:48.627926 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwz2" event={"ID":"b017e67d-043f-4fc3-b342-6480df334c2b","Type":"ContainerDied","Data":"d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983"} Sep 29 17:22:49 crc kubenswrapper[4667]: I0929 17:22:49.362887 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:49 crc kubenswrapper[4667]: I0929 17:22:49.363084 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:49 crc kubenswrapper[4667]: I0929 17:22:49.394475 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:49 crc kubenswrapper[4667]: I0929 17:22:49.645877 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwz2" event={"ID":"b017e67d-043f-4fc3-b342-6480df334c2b","Type":"ContainerStarted","Data":"94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f"} Sep 29 17:22:49 crc kubenswrapper[4667]: I0929 17:22:49.662247 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kfwz2" podStartSLOduration=3.205453939 podStartE2EDuration="5.66221326s" podCreationTimestamp="2025-09-29 17:22:44 +0000 UTC" firstStartedPulling="2025-09-29 17:22:46.618262384 +0000 UTC m=+795.116109153" lastFinishedPulling="2025-09-29 17:22:49.075021706 +0000 UTC m=+797.572868474" observedRunningTime="2025-09-29 17:22:49.659294727 +0000 UTC m=+798.157141497" watchObservedRunningTime="2025-09-29 17:22:49.66221326 +0000 UTC m=+798.160060019" Sep 29 17:22:49 crc kubenswrapper[4667]: I0929 17:22:49.678128 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:51 crc kubenswrapper[4667]: I0929 17:22:51.660337 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nll2m"] Sep 29 17:22:51 crc kubenswrapper[4667]: I0929 17:22:51.660702 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nll2m" podUID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerName="registry-server" containerID="cri-o://4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7" gracePeriod=2 Sep 29 17:22:51 crc kubenswrapper[4667]: I0929 17:22:51.984518 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.136710 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-utilities\") pod \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.136749 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-catalog-content\") pod \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.136888 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqbxf\" (UniqueName: \"kubernetes.io/projected/a8713a8f-02d7-4b54-aeb7-b9d09d017543-kube-api-access-sqbxf\") pod \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\" (UID: \"a8713a8f-02d7-4b54-aeb7-b9d09d017543\") " Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.137397 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-utilities" (OuterVolumeSpecName: "utilities") pod "a8713a8f-02d7-4b54-aeb7-b9d09d017543" (UID: "a8713a8f-02d7-4b54-aeb7-b9d09d017543"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.141051 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8713a8f-02d7-4b54-aeb7-b9d09d017543-kube-api-access-sqbxf" (OuterVolumeSpecName: "kube-api-access-sqbxf") pod "a8713a8f-02d7-4b54-aeb7-b9d09d017543" (UID: "a8713a8f-02d7-4b54-aeb7-b9d09d017543"). InnerVolumeSpecName "kube-api-access-sqbxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.239048 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.239092 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqbxf\" (UniqueName: \"kubernetes.io/projected/a8713a8f-02d7-4b54-aeb7-b9d09d017543-kube-api-access-sqbxf\") on node \"crc\" DevicePath \"\"" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.439706 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8713a8f-02d7-4b54-aeb7-b9d09d017543" (UID: "a8713a8f-02d7-4b54-aeb7-b9d09d017543"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.441710 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8713a8f-02d7-4b54-aeb7-b9d09d017543-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.660908 4667 generic.go:334] "Generic (PLEG): container finished" podID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerID="4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7" exitCode=0 Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.660987 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nll2m" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.661005 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nll2m" event={"ID":"a8713a8f-02d7-4b54-aeb7-b9d09d017543","Type":"ContainerDied","Data":"4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7"} Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.661259 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nll2m" event={"ID":"a8713a8f-02d7-4b54-aeb7-b9d09d017543","Type":"ContainerDied","Data":"331a6dfa36396b179b8127f524c2a0c2bb3701bb8b0906e938a28391ab17b244"} Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.661283 4667 scope.go:117] "RemoveContainer" containerID="4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.677648 4667 scope.go:117] "RemoveContainer" containerID="16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.683059 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nll2m"] Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.686954 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nll2m"] Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.706677 4667 scope.go:117] "RemoveContainer" containerID="8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.718723 4667 scope.go:117] "RemoveContainer" containerID="4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7" Sep 29 17:22:52 crc kubenswrapper[4667]: E0929 17:22:52.719016 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7\": container with ID starting with 4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7 not found: ID does not exist" containerID="4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.719052 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7"} err="failed to get container status \"4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7\": rpc error: code = NotFound desc = could not find container \"4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7\": container with ID starting with 4ef399b37c9c18023bf03dd79ba835129c483da7e17d6681f5927a56e33961b7 not found: ID does not exist" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.719072 4667 scope.go:117] "RemoveContainer" containerID="16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc" Sep 29 17:22:52 crc kubenswrapper[4667]: E0929 17:22:52.719327 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc\": container with ID starting with 16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc not found: ID does not exist" containerID="16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.719359 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc"} err="failed to get container status \"16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc\": rpc error: code = NotFound desc = could not find container \"16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc\": container with ID starting with 16e3c348f27f3d702fa3dc6e185d22e34974f095b32ff052be8de735076ff9dc not found: ID does not exist" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.719383 4667 scope.go:117] "RemoveContainer" containerID="8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8" Sep 29 17:22:52 crc kubenswrapper[4667]: E0929 17:22:52.719642 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8\": container with ID starting with 8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8 not found: ID does not exist" containerID="8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8" Sep 29 17:22:52 crc kubenswrapper[4667]: I0929 17:22:52.719673 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8"} err="failed to get container status \"8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8\": rpc error: code = NotFound desc = could not find container \"8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8\": container with ID starting with 8813306de0ece9a6b438a55bfd27f0f4193645ed69221ddf8e87622a9d05dac8 not found: ID does not exist" Sep 29 17:22:53 crc kubenswrapper[4667]: I0929 17:22:53.825432 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" path="/var/lib/kubelet/pods/a8713a8f-02d7-4b54-aeb7-b9d09d017543/volumes" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.662425 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r9vd2"] Sep 29 17:22:54 crc kubenswrapper[4667]: E0929 17:22:54.662854 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerName="registry-server" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.662867 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerName="registry-server" Sep 29 17:22:54 crc kubenswrapper[4667]: E0929 17:22:54.662879 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerName="extract-content" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.662885 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerName="extract-content" Sep 29 17:22:54 crc kubenswrapper[4667]: E0929 17:22:54.662892 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerName="extract-utilities" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.662898 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerName="extract-utilities" Sep 29 17:22:54 crc kubenswrapper[4667]: E0929 17:22:54.662919 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerName="extract-content" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.662924 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerName="extract-content" Sep 29 17:22:54 crc kubenswrapper[4667]: E0929 17:22:54.662930 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerName="registry-server" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.662935 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerName="registry-server" Sep 29 17:22:54 crc kubenswrapper[4667]: E0929 17:22:54.662944 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerName="extract-utilities" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.662949 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerName="extract-utilities" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.663067 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8713a8f-02d7-4b54-aeb7-b9d09d017543" containerName="registry-server" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.663083 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d3eaf62-b0e0-439e-b085-86af16c929b4" containerName="registry-server" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.663904 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.672610 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r9vd2"] Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.769640 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-catalog-content\") pod \"community-operators-r9vd2\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.769721 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsjtk\" (UniqueName: \"kubernetes.io/projected/bda9390a-1d12-430f-94d9-89117e002b65-kube-api-access-bsjtk\") pod \"community-operators-r9vd2\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.769920 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-utilities\") pod \"community-operators-r9vd2\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.794542 4667 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.794587 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="658afe18-7fa3-4e33-bf5c-cfab27925cb4" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.870992 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-utilities\") pod \"community-operators-r9vd2\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.871196 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-catalog-content\") pod \"community-operators-r9vd2\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.871262 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsjtk\" (UniqueName: \"kubernetes.io/projected/bda9390a-1d12-430f-94d9-89117e002b65-kube-api-access-bsjtk\") pod \"community-operators-r9vd2\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.871453 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-utilities\") pod \"community-operators-r9vd2\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.871584 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-catalog-content\") pod \"community-operators-r9vd2\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.891962 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsjtk\" (UniqueName: \"kubernetes.io/projected/bda9390a-1d12-430f-94d9-89117e002b65-kube-api-access-bsjtk\") pod \"community-operators-r9vd2\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.976614 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.980447 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:54 crc kubenswrapper[4667]: I0929 17:22:54.980487 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:55 crc kubenswrapper[4667]: I0929 17:22:55.014262 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:55 crc kubenswrapper[4667]: I0929 17:22:55.327861 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r9vd2"] Sep 29 17:22:55 crc kubenswrapper[4667]: I0929 17:22:55.676342 4667 generic.go:334] "Generic (PLEG): container finished" podID="bda9390a-1d12-430f-94d9-89117e002b65" containerID="3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276" exitCode=0 Sep 29 17:22:55 crc kubenswrapper[4667]: I0929 17:22:55.676418 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vd2" event={"ID":"bda9390a-1d12-430f-94d9-89117e002b65","Type":"ContainerDied","Data":"3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276"} Sep 29 17:22:55 crc kubenswrapper[4667]: I0929 17:22:55.676470 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vd2" event={"ID":"bda9390a-1d12-430f-94d9-89117e002b65","Type":"ContainerStarted","Data":"64ef2e50cb77a9c48008ea34549b34476101c4efc61cf87df728c6e36fc5b767"} Sep 29 17:22:55 crc kubenswrapper[4667]: I0929 17:22:55.706155 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:56 crc kubenswrapper[4667]: I0929 17:22:56.682310 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vd2" event={"ID":"bda9390a-1d12-430f-94d9-89117e002b65","Type":"ContainerStarted","Data":"e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6"} Sep 29 17:22:56 crc kubenswrapper[4667]: I0929 17:22:56.897051 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:22:56 crc kubenswrapper[4667]: I0929 17:22:56.897269 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:22:57 crc kubenswrapper[4667]: I0929 17:22:57.688878 4667 generic.go:334] "Generic (PLEG): container finished" podID="bda9390a-1d12-430f-94d9-89117e002b65" containerID="e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6" exitCode=0 Sep 29 17:22:57 crc kubenswrapper[4667]: I0929 17:22:57.688918 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vd2" event={"ID":"bda9390a-1d12-430f-94d9-89117e002b65","Type":"ContainerDied","Data":"e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6"} Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.061501 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfwz2"] Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.061699 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kfwz2" podUID="b017e67d-043f-4fc3-b342-6480df334c2b" containerName="registry-server" containerID="cri-o://94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f" gracePeriod=2 Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.418588 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.524216 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-utilities\") pod \"b017e67d-043f-4fc3-b342-6480df334c2b\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.524344 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-catalog-content\") pod \"b017e67d-043f-4fc3-b342-6480df334c2b\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.524394 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r22h\" (UniqueName: \"kubernetes.io/projected/b017e67d-043f-4fc3-b342-6480df334c2b-kube-api-access-8r22h\") pod \"b017e67d-043f-4fc3-b342-6480df334c2b\" (UID: \"b017e67d-043f-4fc3-b342-6480df334c2b\") " Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.524907 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-utilities" (OuterVolumeSpecName: "utilities") pod "b017e67d-043f-4fc3-b342-6480df334c2b" (UID: "b017e67d-043f-4fc3-b342-6480df334c2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.528329 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b017e67d-043f-4fc3-b342-6480df334c2b-kube-api-access-8r22h" (OuterVolumeSpecName: "kube-api-access-8r22h") pod "b017e67d-043f-4fc3-b342-6480df334c2b" (UID: "b017e67d-043f-4fc3-b342-6480df334c2b"). InnerVolumeSpecName "kube-api-access-8r22h". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.553986 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b017e67d-043f-4fc3-b342-6480df334c2b" (UID: "b017e67d-043f-4fc3-b342-6480df334c2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.626437 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.626460 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r22h\" (UniqueName: \"kubernetes.io/projected/b017e67d-043f-4fc3-b342-6480df334c2b-kube-api-access-8r22h\") on node \"crc\" DevicePath \"\"" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.626488 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b017e67d-043f-4fc3-b342-6480df334c2b-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.696269 4667 generic.go:334] "Generic (PLEG): container finished" podID="b017e67d-043f-4fc3-b342-6480df334c2b" containerID="94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f" exitCode=0 Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.696328 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfwz2" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.696333 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwz2" event={"ID":"b017e67d-043f-4fc3-b342-6480df334c2b","Type":"ContainerDied","Data":"94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f"} Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.696553 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfwz2" event={"ID":"b017e67d-043f-4fc3-b342-6480df334c2b","Type":"ContainerDied","Data":"36481a59d3509185688dc305c9b793d7702e0ff6df7e29fa58671810b39ca288"} Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.696577 4667 scope.go:117] "RemoveContainer" containerID="94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.699386 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vd2" event={"ID":"bda9390a-1d12-430f-94d9-89117e002b65","Type":"ContainerStarted","Data":"0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683"} Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.708920 4667 scope.go:117] "RemoveContainer" containerID="d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.715517 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r9vd2" podStartSLOduration=2.044328806 podStartE2EDuration="4.715506979s" podCreationTimestamp="2025-09-29 17:22:54 +0000 UTC" firstStartedPulling="2025-09-29 17:22:55.677529153 +0000 UTC m=+804.175375923" lastFinishedPulling="2025-09-29 17:22:58.348707327 +0000 UTC m=+806.846554096" observedRunningTime="2025-09-29 17:22:58.713279529 +0000 UTC m=+807.211126298" watchObservedRunningTime="2025-09-29 17:22:58.715506979 +0000 UTC m=+807.213353748" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.722591 4667 scope.go:117] "RemoveContainer" containerID="6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.724859 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfwz2"] Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.729615 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kfwz2"] Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.740728 4667 scope.go:117] "RemoveContainer" containerID="94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f" Sep 29 17:22:58 crc kubenswrapper[4667]: E0929 17:22:58.741092 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f\": container with ID starting with 94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f not found: ID does not exist" containerID="94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.741141 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f"} err="failed to get container status \"94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f\": rpc error: code = NotFound desc = could not find container \"94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f\": container with ID starting with 94db0801bfda3fc562ba54c19df41fe4f001f1c1e91253bb3f3cfe14eb268a0f not found: ID does not exist" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.741162 4667 scope.go:117] "RemoveContainer" containerID="d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983" Sep 29 17:22:58 crc kubenswrapper[4667]: E0929 17:22:58.741465 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983\": container with ID starting with d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983 not found: ID does not exist" containerID="d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.741485 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983"} err="failed to get container status \"d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983\": rpc error: code = NotFound desc = could not find container \"d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983\": container with ID starting with d35f93a4298517a54eca72fa79932eff107957172729093acfc0eb1bd9013983 not found: ID does not exist" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.741521 4667 scope.go:117] "RemoveContainer" containerID="6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a" Sep 29 17:22:58 crc kubenswrapper[4667]: E0929 17:22:58.741712 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a\": container with ID starting with 6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a not found: ID does not exist" containerID="6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a" Sep 29 17:22:58 crc kubenswrapper[4667]: I0929 17:22:58.741733 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a"} err="failed to get container status \"6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a\": rpc error: code = NotFound desc = could not find container \"6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a\": container with ID starting with 6d8a3fd64e4c07434edebd55b26f7d3397b7b511a29654fccc26251526f7b23a not found: ID does not exist" Sep 29 17:22:59 crc kubenswrapper[4667]: I0929 17:22:59.821551 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b017e67d-043f-4fc3-b342-6480df334c2b" path="/var/lib/kubelet/pods/b017e67d-043f-4fc3-b342-6480df334c2b/volumes" Sep 29 17:23:04 crc kubenswrapper[4667]: I0929 17:23:04.794724 4667 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Sep 29 17:23:04 crc kubenswrapper[4667]: I0929 17:23:04.794967 4667 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="658afe18-7fa3-4e33-bf5c-cfab27925cb4" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Sep 29 17:23:04 crc kubenswrapper[4667]: I0929 17:23:04.977556 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:23:04 crc kubenswrapper[4667]: I0929 17:23:04.977610 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:23:05 crc kubenswrapper[4667]: I0929 17:23:05.007604 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:23:05 crc kubenswrapper[4667]: I0929 17:23:05.758907 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:23:05 crc kubenswrapper[4667]: I0929 17:23:05.789044 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r9vd2"] Sep 29 17:23:07 crc kubenswrapper[4667]: I0929 17:23:07.740507 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r9vd2" podUID="bda9390a-1d12-430f-94d9-89117e002b65" containerName="registry-server" containerID="cri-o://0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683" gracePeriod=2 Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.068547 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.148417 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-catalog-content\") pod \"bda9390a-1d12-430f-94d9-89117e002b65\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.148481 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-utilities\") pod \"bda9390a-1d12-430f-94d9-89117e002b65\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.149162 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-utilities" (OuterVolumeSpecName: "utilities") pod "bda9390a-1d12-430f-94d9-89117e002b65" (UID: "bda9390a-1d12-430f-94d9-89117e002b65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.149200 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsjtk\" (UniqueName: \"kubernetes.io/projected/bda9390a-1d12-430f-94d9-89117e002b65-kube-api-access-bsjtk\") pod \"bda9390a-1d12-430f-94d9-89117e002b65\" (UID: \"bda9390a-1d12-430f-94d9-89117e002b65\") " Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.149520 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.152891 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bda9390a-1d12-430f-94d9-89117e002b65-kube-api-access-bsjtk" (OuterVolumeSpecName: "kube-api-access-bsjtk") pod "bda9390a-1d12-430f-94d9-89117e002b65" (UID: "bda9390a-1d12-430f-94d9-89117e002b65"). InnerVolumeSpecName "kube-api-access-bsjtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.182874 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bda9390a-1d12-430f-94d9-89117e002b65" (UID: "bda9390a-1d12-430f-94d9-89117e002b65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.250328 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsjtk\" (UniqueName: \"kubernetes.io/projected/bda9390a-1d12-430f-94d9-89117e002b65-kube-api-access-bsjtk\") on node \"crc\" DevicePath \"\"" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.250355 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bda9390a-1d12-430f-94d9-89117e002b65-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.749369 4667 generic.go:334] "Generic (PLEG): container finished" podID="bda9390a-1d12-430f-94d9-89117e002b65" containerID="0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683" exitCode=0 Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.749411 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9vd2" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.749430 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vd2" event={"ID":"bda9390a-1d12-430f-94d9-89117e002b65","Type":"ContainerDied","Data":"0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683"} Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.749778 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vd2" event={"ID":"bda9390a-1d12-430f-94d9-89117e002b65","Type":"ContainerDied","Data":"64ef2e50cb77a9c48008ea34549b34476101c4efc61cf87df728c6e36fc5b767"} Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.749802 4667 scope.go:117] "RemoveContainer" containerID="0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.766066 4667 scope.go:117] "RemoveContainer" containerID="e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.772598 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r9vd2"] Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.776560 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r9vd2"] Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.800172 4667 scope.go:117] "RemoveContainer" containerID="3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.812217 4667 scope.go:117] "RemoveContainer" containerID="0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683" Sep 29 17:23:08 crc kubenswrapper[4667]: E0929 17:23:08.812522 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683\": container with ID starting with 0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683 not found: ID does not exist" containerID="0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.812558 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683"} err="failed to get container status \"0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683\": rpc error: code = NotFound desc = could not find container \"0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683\": container with ID starting with 0e6dcffeab25518caca0e826bbf1edb9a9b592dd81d9c1038bc695855e0eb683 not found: ID does not exist" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.812580 4667 scope.go:117] "RemoveContainer" containerID="e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6" Sep 29 17:23:08 crc kubenswrapper[4667]: E0929 17:23:08.812920 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6\": container with ID starting with e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6 not found: ID does not exist" containerID="e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.812955 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6"} err="failed to get container status \"e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6\": rpc error: code = NotFound desc = could not find container \"e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6\": container with ID starting with e31030a447cc4e5015f62554d42c9f192497e10ea7f216d8c1af0cf2e1f4bab6 not found: ID does not exist" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.812978 4667 scope.go:117] "RemoveContainer" containerID="3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276" Sep 29 17:23:08 crc kubenswrapper[4667]: E0929 17:23:08.813213 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276\": container with ID starting with 3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276 not found: ID does not exist" containerID="3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276" Sep 29 17:23:08 crc kubenswrapper[4667]: I0929 17:23:08.813235 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276"} err="failed to get container status \"3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276\": rpc error: code = NotFound desc = could not find container \"3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276\": container with ID starting with 3f327bb4143084fc954d559576a780a664a366172c6a366f1f0c57b197013276 not found: ID does not exist" Sep 29 17:23:09 crc kubenswrapper[4667]: I0929 17:23:09.823117 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bda9390a-1d12-430f-94d9-89117e002b65" path="/var/lib/kubelet/pods/bda9390a-1d12-430f-94d9-89117e002b65/volumes" Sep 29 17:23:14 crc kubenswrapper[4667]: I0929 17:23:14.794760 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Sep 29 17:23:26 crc kubenswrapper[4667]: I0929 17:23:26.896979 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:23:26 crc kubenswrapper[4667]: I0929 17:23:26.897487 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:23:26 crc kubenswrapper[4667]: I0929 17:23:26.897522 4667 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:23:26 crc kubenswrapper[4667]: I0929 17:23:26.897931 4667 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24409b364de3d67a164134cbdd70a1ed09fbd088e081c52528c8d39cd38525a9"} pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 17:23:26 crc kubenswrapper[4667]: I0929 17:23:26.897981 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" containerID="cri-o://24409b364de3d67a164134cbdd70a1ed09fbd088e081c52528c8d39cd38525a9" gracePeriod=600 Sep 29 17:23:27 crc kubenswrapper[4667]: I0929 17:23:27.856661 4667 generic.go:334] "Generic (PLEG): container finished" podID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerID="24409b364de3d67a164134cbdd70a1ed09fbd088e081c52528c8d39cd38525a9" exitCode=0 Sep 29 17:23:27 crc kubenswrapper[4667]: I0929 17:23:27.856728 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerDied","Data":"24409b364de3d67a164134cbdd70a1ed09fbd088e081c52528c8d39cd38525a9"} Sep 29 17:23:27 crc kubenswrapper[4667]: I0929 17:23:27.857035 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"394308b51f762322854d3b0cdd5e2f611129dcf7d5f915f00008534ba0f925d4"} Sep 29 17:23:27 crc kubenswrapper[4667]: I0929 17:23:27.857056 4667 scope.go:117] "RemoveContainer" containerID="2b7b6f639c7d54c34cc1faa704af3d81c5333b8dbbef3163c511e28b934a21eb" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.036316 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-wvmfl"] Sep 29 17:23:34 crc kubenswrapper[4667]: E0929 17:23:34.037420 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda9390a-1d12-430f-94d9-89117e002b65" containerName="extract-content" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.037441 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda9390a-1d12-430f-94d9-89117e002b65" containerName="extract-content" Sep 29 17:23:34 crc kubenswrapper[4667]: E0929 17:23:34.037461 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b017e67d-043f-4fc3-b342-6480df334c2b" containerName="extract-content" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.037467 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="b017e67d-043f-4fc3-b342-6480df334c2b" containerName="extract-content" Sep 29 17:23:34 crc kubenswrapper[4667]: E0929 17:23:34.037477 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b017e67d-043f-4fc3-b342-6480df334c2b" containerName="extract-utilities" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.037485 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="b017e67d-043f-4fc3-b342-6480df334c2b" containerName="extract-utilities" Sep 29 17:23:34 crc kubenswrapper[4667]: E0929 17:23:34.037493 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda9390a-1d12-430f-94d9-89117e002b65" containerName="extract-utilities" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.037499 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda9390a-1d12-430f-94d9-89117e002b65" containerName="extract-utilities" Sep 29 17:23:34 crc kubenswrapper[4667]: E0929 17:23:34.037526 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda9390a-1d12-430f-94d9-89117e002b65" containerName="registry-server" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.037533 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda9390a-1d12-430f-94d9-89117e002b65" containerName="registry-server" Sep 29 17:23:34 crc kubenswrapper[4667]: E0929 17:23:34.037541 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b017e67d-043f-4fc3-b342-6480df334c2b" containerName="registry-server" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.037549 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="b017e67d-043f-4fc3-b342-6480df334c2b" containerName="registry-server" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.037711 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="bda9390a-1d12-430f-94d9-89117e002b65" containerName="registry-server" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.037727 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="b017e67d-043f-4fc3-b342-6480df334c2b" containerName="registry-server" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.038561 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.043124 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.043292 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.043132 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.043458 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.043630 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-bt728" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.056312 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.064293 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-wvmfl"] Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096348 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-sa-token\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096458 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-datadir\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096547 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kklk\" (UniqueName: \"kubernetes.io/projected/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-kube-api-access-5kklk\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096578 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-collector-token\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096622 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-metrics\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096701 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-trusted-ca\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096760 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-tmp\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096816 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-config-openshift-service-cacrt\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096838 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-collector-syslog-receiver\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.096979 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-config\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.097025 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-entrypoint\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.199236 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-sa-token\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.199474 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-datadir\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.199596 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kklk\" (UniqueName: \"kubernetes.io/projected/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-kube-api-access-5kklk\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.199707 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-collector-token\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.199806 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-metrics\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.199909 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-trusted-ca\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.200021 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-tmp\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.200124 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-config-openshift-service-cacrt\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.199737 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-datadir\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.200225 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-collector-syslog-receiver\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.200406 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-config\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.200441 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-entrypoint\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.200883 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-trusted-ca\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.200896 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-config-openshift-service-cacrt\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.201091 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-config\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.201164 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-entrypoint\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.205943 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-collector-syslog-receiver\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.205974 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-metrics\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.214144 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-collector-token\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.215330 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-sa-token\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.215340 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-tmp\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.229163 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kklk\" (UniqueName: \"kubernetes.io/projected/4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a-kube-api-access-5kklk\") pod \"collector-wvmfl\" (UID: \"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a\") " pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.363273 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-wvmfl" Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.736526 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-wvmfl"] Sep 29 17:23:34 crc kubenswrapper[4667]: I0929 17:23:34.899226 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-wvmfl" event={"ID":"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a","Type":"ContainerStarted","Data":"738fd77c1fbd37e1673a4ba6779a6757c85ddb72bfa7d59affb0e3ac21bb16b7"} Sep 29 17:23:39 crc kubenswrapper[4667]: I0929 17:23:39.932091 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-wvmfl" event={"ID":"4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a","Type":"ContainerStarted","Data":"43e561697d06a5022a82c4e716d3b99db6cb3d815f1131f18e8c55b0fc98f94a"} Sep 29 17:23:39 crc kubenswrapper[4667]: I0929 17:23:39.949016 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-wvmfl" podStartSLOduration=1.012271629 podStartE2EDuration="5.948987812s" podCreationTimestamp="2025-09-29 17:23:34 +0000 UTC" firstStartedPulling="2025-09-29 17:23:34.745466981 +0000 UTC m=+843.243313750" lastFinishedPulling="2025-09-29 17:23:39.682183163 +0000 UTC m=+848.180029933" observedRunningTime="2025-09-29 17:23:39.948364356 +0000 UTC m=+848.446211125" watchObservedRunningTime="2025-09-29 17:23:39.948987812 +0000 UTC m=+848.446834581" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.626956 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll"] Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.628509 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.630980 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.641815 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll"] Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.653473 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.653733 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snwgq\" (UniqueName: \"kubernetes.io/projected/19f4040e-c856-4d41-a9a0-977dbce39c98-kube-api-access-snwgq\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.653770 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.755333 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snwgq\" (UniqueName: \"kubernetes.io/projected/19f4040e-c856-4d41-a9a0-977dbce39c98-kube-api-access-snwgq\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.755408 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.755458 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.755987 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-bundle\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.756134 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-util\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.775552 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snwgq\" (UniqueName: \"kubernetes.io/projected/19f4040e-c856-4d41-a9a0-977dbce39c98-kube-api-access-snwgq\") pod \"9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:51 crc kubenswrapper[4667]: I0929 17:23:51.942311 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:52 crc kubenswrapper[4667]: I0929 17:23:52.317319 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll"] Sep 29 17:23:53 crc kubenswrapper[4667]: I0929 17:23:53.006635 4667 generic.go:334] "Generic (PLEG): container finished" podID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerID="c87c03707c0fed7c2322cf9542eb7b5a37399a3b26fea66090897d5063cd9970" exitCode=0 Sep 29 17:23:53 crc kubenswrapper[4667]: I0929 17:23:53.006721 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" event={"ID":"19f4040e-c856-4d41-a9a0-977dbce39c98","Type":"ContainerDied","Data":"c87c03707c0fed7c2322cf9542eb7b5a37399a3b26fea66090897d5063cd9970"} Sep 29 17:23:53 crc kubenswrapper[4667]: I0929 17:23:53.007015 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" event={"ID":"19f4040e-c856-4d41-a9a0-977dbce39c98","Type":"ContainerStarted","Data":"afdddc994dee5bc3634bdb1706318c638099d2b1a3806e2f68ca155b029f8ac9"} Sep 29 17:23:55 crc kubenswrapper[4667]: I0929 17:23:55.020115 4667 generic.go:334] "Generic (PLEG): container finished" podID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerID="02366599b47126f01d2a035c6122c0cc3d7127e13e7418c602401875215ce52f" exitCode=0 Sep 29 17:23:55 crc kubenswrapper[4667]: I0929 17:23:55.020195 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" event={"ID":"19f4040e-c856-4d41-a9a0-977dbce39c98","Type":"ContainerDied","Data":"02366599b47126f01d2a035c6122c0cc3d7127e13e7418c602401875215ce52f"} Sep 29 17:23:56 crc kubenswrapper[4667]: I0929 17:23:56.027513 4667 generic.go:334] "Generic (PLEG): container finished" podID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerID="6d68b0789ea53659d83ea56847b2ea7311a73c8d968b053449c8c5d4d578373b" exitCode=0 Sep 29 17:23:56 crc kubenswrapper[4667]: I0929 17:23:56.027558 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" event={"ID":"19f4040e-c856-4d41-a9a0-977dbce39c98","Type":"ContainerDied","Data":"6d68b0789ea53659d83ea56847b2ea7311a73c8d968b053449c8c5d4d578373b"} Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.278635 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.326830 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-util\") pod \"19f4040e-c856-4d41-a9a0-977dbce39c98\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.326951 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-bundle\") pod \"19f4040e-c856-4d41-a9a0-977dbce39c98\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.327007 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snwgq\" (UniqueName: \"kubernetes.io/projected/19f4040e-c856-4d41-a9a0-977dbce39c98-kube-api-access-snwgq\") pod \"19f4040e-c856-4d41-a9a0-977dbce39c98\" (UID: \"19f4040e-c856-4d41-a9a0-977dbce39c98\") " Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.327395 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-bundle" (OuterVolumeSpecName: "bundle") pod "19f4040e-c856-4d41-a9a0-977dbce39c98" (UID: "19f4040e-c856-4d41-a9a0-977dbce39c98"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.327564 4667 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.334932 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f4040e-c856-4d41-a9a0-977dbce39c98-kube-api-access-snwgq" (OuterVolumeSpecName: "kube-api-access-snwgq") pod "19f4040e-c856-4d41-a9a0-977dbce39c98" (UID: "19f4040e-c856-4d41-a9a0-977dbce39c98"). InnerVolumeSpecName "kube-api-access-snwgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.343988 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-util" (OuterVolumeSpecName: "util") pod "19f4040e-c856-4d41-a9a0-977dbce39c98" (UID: "19f4040e-c856-4d41-a9a0-977dbce39c98"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.428943 4667 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/19f4040e-c856-4d41-a9a0-977dbce39c98-util\") on node \"crc\" DevicePath \"\"" Sep 29 17:23:57 crc kubenswrapper[4667]: I0929 17:23:57.428969 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snwgq\" (UniqueName: \"kubernetes.io/projected/19f4040e-c856-4d41-a9a0-977dbce39c98-kube-api-access-snwgq\") on node \"crc\" DevicePath \"\"" Sep 29 17:23:58 crc kubenswrapper[4667]: I0929 17:23:58.038268 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" event={"ID":"19f4040e-c856-4d41-a9a0-977dbce39c98","Type":"ContainerDied","Data":"afdddc994dee5bc3634bdb1706318c638099d2b1a3806e2f68ca155b029f8ac9"} Sep 29 17:23:58 crc kubenswrapper[4667]: I0929 17:23:58.038304 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afdddc994dee5bc3634bdb1706318c638099d2b1a3806e2f68ca155b029f8ac9" Sep 29 17:23:58 crc kubenswrapper[4667]: I0929 17:23:58.038323 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.265287 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4"] Sep 29 17:24:03 crc kubenswrapper[4667]: E0929 17:24:03.266180 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerName="pull" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.266198 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerName="pull" Sep 29 17:24:03 crc kubenswrapper[4667]: E0929 17:24:03.266226 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerName="extract" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.266231 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerName="extract" Sep 29 17:24:03 crc kubenswrapper[4667]: E0929 17:24:03.266244 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerName="util" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.266250 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerName="util" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.266410 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="19f4040e-c856-4d41-a9a0-977dbce39c98" containerName="extract" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.266986 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.268672 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4qf4n" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.269137 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.269461 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.283811 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4"] Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.412868 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptwgh\" (UniqueName: \"kubernetes.io/projected/6d4c734e-b4e1-4daf-ba5f-24d5a4e2c163-kube-api-access-ptwgh\") pod \"nmstate-operator-5d6f6cfd66-k77n4\" (UID: \"6d4c734e-b4e1-4daf-ba5f-24d5a4e2c163\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.514887 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptwgh\" (UniqueName: \"kubernetes.io/projected/6d4c734e-b4e1-4daf-ba5f-24d5a4e2c163-kube-api-access-ptwgh\") pod \"nmstate-operator-5d6f6cfd66-k77n4\" (UID: \"6d4c734e-b4e1-4daf-ba5f-24d5a4e2c163\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.530544 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptwgh\" (UniqueName: \"kubernetes.io/projected/6d4c734e-b4e1-4daf-ba5f-24d5a4e2c163-kube-api-access-ptwgh\") pod \"nmstate-operator-5d6f6cfd66-k77n4\" (UID: \"6d4c734e-b4e1-4daf-ba5f-24d5a4e2c163\") " pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.580563 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4" Sep 29 17:24:03 crc kubenswrapper[4667]: I0929 17:24:03.945974 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4"] Sep 29 17:24:04 crc kubenswrapper[4667]: I0929 17:24:04.077681 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4" event={"ID":"6d4c734e-b4e1-4daf-ba5f-24d5a4e2c163","Type":"ContainerStarted","Data":"160467b7c29388049b68eb37589125d25e32b68d4aba26e07f3eeaec1512c7bf"} Sep 29 17:24:07 crc kubenswrapper[4667]: I0929 17:24:07.096477 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4" event={"ID":"6d4c734e-b4e1-4daf-ba5f-24d5a4e2c163","Type":"ContainerStarted","Data":"33b3c0a0e06056cbd54354a6144a1bb13fc79af48447a659dfa987b1dc3c41fb"} Sep 29 17:24:07 crc kubenswrapper[4667]: I0929 17:24:07.112992 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5d6f6cfd66-k77n4" podStartSLOduration=1.768586041 podStartE2EDuration="4.112974997s" podCreationTimestamp="2025-09-29 17:24:03 +0000 UTC" firstStartedPulling="2025-09-29 17:24:03.957859041 +0000 UTC m=+872.455705810" lastFinishedPulling="2025-09-29 17:24:06.302247997 +0000 UTC m=+874.800094766" observedRunningTime="2025-09-29 17:24:07.108073987 +0000 UTC m=+875.605920756" watchObservedRunningTime="2025-09-29 17:24:07.112974997 +0000 UTC m=+875.610821767" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.499247 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg"] Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.500383 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.503167 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-ghvs7" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.509654 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg"] Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.519181 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w"] Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.520259 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.522388 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.525533 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-hcvkt"] Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.526313 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.541479 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w"] Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.559280 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/58bb8f72-34d0-44a3-a094-dc59c23fe49b-nmstate-lock\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.559366 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xslkt\" (UniqueName: \"kubernetes.io/projected/18dfc1a3-add5-49cb-860c-5361e887a24d-kube-api-access-xslkt\") pod \"nmstate-webhook-6d689559c5-h5d7w\" (UID: \"18dfc1a3-add5-49cb-860c-5361e887a24d\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.559425 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/58bb8f72-34d0-44a3-a094-dc59c23fe49b-ovs-socket\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.559462 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxrjd\" (UniqueName: \"kubernetes.io/projected/b76f4f28-f053-48ff-bb61-524505148d8d-kube-api-access-gxrjd\") pod \"nmstate-metrics-58fcddf996-nvhmg\" (UID: \"b76f4f28-f053-48ff-bb61-524505148d8d\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.559481 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/18dfc1a3-add5-49cb-860c-5361e887a24d-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h5d7w\" (UID: \"18dfc1a3-add5-49cb-860c-5361e887a24d\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.559496 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvfcw\" (UniqueName: \"kubernetes.io/projected/58bb8f72-34d0-44a3-a094-dc59c23fe49b-kube-api-access-hvfcw\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.559511 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/58bb8f72-34d0-44a3-a094-dc59c23fe49b-dbus-socket\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.620695 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh"] Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.621477 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.622801 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.622957 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.622999 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-l78qp" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.628786 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh"] Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660618 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/58bb8f72-34d0-44a3-a094-dc59c23fe49b-nmstate-lock\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660693 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4lwt\" (UniqueName: \"kubernetes.io/projected/be8fb545-6b61-4464-b940-0fab07461beb-kube-api-access-v4lwt\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660719 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xslkt\" (UniqueName: \"kubernetes.io/projected/18dfc1a3-add5-49cb-860c-5361e887a24d-kube-api-access-xslkt\") pod \"nmstate-webhook-6d689559c5-h5d7w\" (UID: \"18dfc1a3-add5-49cb-860c-5361e887a24d\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660735 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/58bb8f72-34d0-44a3-a094-dc59c23fe49b-nmstate-lock\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660761 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/58bb8f72-34d0-44a3-a094-dc59c23fe49b-ovs-socket\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660798 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxrjd\" (UniqueName: \"kubernetes.io/projected/b76f4f28-f053-48ff-bb61-524505148d8d-kube-api-access-gxrjd\") pod \"nmstate-metrics-58fcddf996-nvhmg\" (UID: \"b76f4f28-f053-48ff-bb61-524505148d8d\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660814 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/be8fb545-6b61-4464-b940-0fab07461beb-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660834 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/18dfc1a3-add5-49cb-860c-5361e887a24d-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h5d7w\" (UID: \"18dfc1a3-add5-49cb-860c-5361e887a24d\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660868 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/58bb8f72-34d0-44a3-a094-dc59c23fe49b-dbus-socket\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660867 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/58bb8f72-34d0-44a3-a094-dc59c23fe49b-ovs-socket\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660886 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvfcw\" (UniqueName: \"kubernetes.io/projected/58bb8f72-34d0-44a3-a094-dc59c23fe49b-kube-api-access-hvfcw\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.660928 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/be8fb545-6b61-4464-b940-0fab07461beb-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:12 crc kubenswrapper[4667]: E0929 17:24:12.661093 4667 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Sep 29 17:24:12 crc kubenswrapper[4667]: E0929 17:24:12.661139 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18dfc1a3-add5-49cb-860c-5361e887a24d-tls-key-pair podName:18dfc1a3-add5-49cb-860c-5361e887a24d nodeName:}" failed. No retries permitted until 2025-09-29 17:24:13.161125738 +0000 UTC m=+881.658972507 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/18dfc1a3-add5-49cb-860c-5361e887a24d-tls-key-pair") pod "nmstate-webhook-6d689559c5-h5d7w" (UID: "18dfc1a3-add5-49cb-860c-5361e887a24d") : secret "openshift-nmstate-webhook" not found Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.661159 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/58bb8f72-34d0-44a3-a094-dc59c23fe49b-dbus-socket\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.678304 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvfcw\" (UniqueName: \"kubernetes.io/projected/58bb8f72-34d0-44a3-a094-dc59c23fe49b-kube-api-access-hvfcw\") pod \"nmstate-handler-hcvkt\" (UID: \"58bb8f72-34d0-44a3-a094-dc59c23fe49b\") " pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.678475 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xslkt\" (UniqueName: \"kubernetes.io/projected/18dfc1a3-add5-49cb-860c-5361e887a24d-kube-api-access-xslkt\") pod \"nmstate-webhook-6d689559c5-h5d7w\" (UID: \"18dfc1a3-add5-49cb-860c-5361e887a24d\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.678798 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxrjd\" (UniqueName: \"kubernetes.io/projected/b76f4f28-f053-48ff-bb61-524505148d8d-kube-api-access-gxrjd\") pod \"nmstate-metrics-58fcddf996-nvhmg\" (UID: \"b76f4f28-f053-48ff-bb61-524505148d8d\") " pod="openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.762156 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/be8fb545-6b61-4464-b940-0fab07461beb-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.762575 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/be8fb545-6b61-4464-b940-0fab07461beb-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.762714 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4lwt\" (UniqueName: \"kubernetes.io/projected/be8fb545-6b61-4464-b940-0fab07461beb-kube-api-access-v4lwt\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:12 crc kubenswrapper[4667]: E0929 17:24:12.762781 4667 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Sep 29 17:24:12 crc kubenswrapper[4667]: E0929 17:24:12.762889 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be8fb545-6b61-4464-b940-0fab07461beb-plugin-serving-cert podName:be8fb545-6b61-4464-b940-0fab07461beb nodeName:}" failed. No retries permitted until 2025-09-29 17:24:13.262872668 +0000 UTC m=+881.760719437 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/be8fb545-6b61-4464-b940-0fab07461beb-plugin-serving-cert") pod "nmstate-console-plugin-864bb6dfb5-wzvhh" (UID: "be8fb545-6b61-4464-b940-0fab07461beb") : secret "plugin-serving-cert" not found Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.763805 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/be8fb545-6b61-4464-b940-0fab07461beb-nginx-conf\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.772440 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-66c4c8d9c4-x45lt"] Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.773266 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.778291 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4lwt\" (UniqueName: \"kubernetes.io/projected/be8fb545-6b61-4464-b940-0fab07461beb-kube-api-access-v4lwt\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.785910 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-66c4c8d9c4-x45lt"] Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.814499 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.844888 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.864304 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2791e00e-ed2b-4c22-9cc1-43b616beba71-console-serving-cert\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.864347 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-console-config\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.864419 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqpck\" (UniqueName: \"kubernetes.io/projected/2791e00e-ed2b-4c22-9cc1-43b616beba71-kube-api-access-dqpck\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.864449 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-service-ca\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.864555 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-oauth-serving-cert\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.864581 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2791e00e-ed2b-4c22-9cc1-43b616beba71-console-oauth-config\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.864649 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-trusted-ca-bundle\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.966378 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-console-config\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.966684 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqpck\" (UniqueName: \"kubernetes.io/projected/2791e00e-ed2b-4c22-9cc1-43b616beba71-kube-api-access-dqpck\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.966719 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-service-ca\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.966806 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-oauth-serving-cert\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.966829 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2791e00e-ed2b-4c22-9cc1-43b616beba71-console-oauth-config\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.966900 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-trusted-ca-bundle\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.967005 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2791e00e-ed2b-4c22-9cc1-43b616beba71-console-serving-cert\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.967555 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-oauth-serving-cert\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.968080 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-service-ca\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.971523 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-trusted-ca-bundle\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.971950 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2791e00e-ed2b-4c22-9cc1-43b616beba71-console-config\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.972069 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2791e00e-ed2b-4c22-9cc1-43b616beba71-console-serving-cert\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.973385 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2791e00e-ed2b-4c22-9cc1-43b616beba71-console-oauth-config\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:12 crc kubenswrapper[4667]: I0929 17:24:12.980338 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqpck\" (UniqueName: \"kubernetes.io/projected/2791e00e-ed2b-4c22-9cc1-43b616beba71-kube-api-access-dqpck\") pod \"console-66c4c8d9c4-x45lt\" (UID: \"2791e00e-ed2b-4c22-9cc1-43b616beba71\") " pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.107765 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.125527 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hcvkt" event={"ID":"58bb8f72-34d0-44a3-a094-dc59c23fe49b","Type":"ContainerStarted","Data":"2b872acc4a873a75a317ce8497fd48323f1cdf796459836e6cc9684271e32eff"} Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.169792 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/18dfc1a3-add5-49cb-860c-5361e887a24d-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h5d7w\" (UID: \"18dfc1a3-add5-49cb-860c-5361e887a24d\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.173337 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/18dfc1a3-add5-49cb-860c-5361e887a24d-tls-key-pair\") pod \"nmstate-webhook-6d689559c5-h5d7w\" (UID: \"18dfc1a3-add5-49cb-860c-5361e887a24d\") " pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.221533 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg"] Sep 29 17:24:13 crc kubenswrapper[4667]: W0929 17:24:13.226887 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb76f4f28_f053_48ff_bb61_524505148d8d.slice/crio-5dd6c25b2123c38f6fbb7df5da61b018357d4dcf2ae6abaa360cd39299eacd1e WatchSource:0}: Error finding container 5dd6c25b2123c38f6fbb7df5da61b018357d4dcf2ae6abaa360cd39299eacd1e: Status 404 returned error can't find the container with id 5dd6c25b2123c38f6fbb7df5da61b018357d4dcf2ae6abaa360cd39299eacd1e Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.271582 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/be8fb545-6b61-4464-b940-0fab07461beb-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.274249 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/be8fb545-6b61-4464-b940-0fab07461beb-plugin-serving-cert\") pod \"nmstate-console-plugin-864bb6dfb5-wzvhh\" (UID: \"be8fb545-6b61-4464-b940-0fab07461beb\") " pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.435330 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.455994 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-66c4c8d9c4-x45lt"] Sep 29 17:24:13 crc kubenswrapper[4667]: W0929 17:24:13.458731 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2791e00e_ed2b_4c22_9cc1_43b616beba71.slice/crio-4b6f04e116922ef78313984ed3a1575bc41f4a3efc7465ab423ba786f4716783 WatchSource:0}: Error finding container 4b6f04e116922ef78313984ed3a1575bc41f4a3efc7465ab423ba786f4716783: Status 404 returned error can't find the container with id 4b6f04e116922ef78313984ed3a1575bc41f4a3efc7465ab423ba786f4716783 Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.533414 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.785807 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh"] Sep 29 17:24:13 crc kubenswrapper[4667]: W0929 17:24:13.797114 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe8fb545_6b61_4464_b940_0fab07461beb.slice/crio-c2d162adafad4aedcd23f0f9f6c222e03ef9b5e7a202b40b6c7d9b55d701e754 WatchSource:0}: Error finding container c2d162adafad4aedcd23f0f9f6c222e03ef9b5e7a202b40b6c7d9b55d701e754: Status 404 returned error can't find the container with id c2d162adafad4aedcd23f0f9f6c222e03ef9b5e7a202b40b6c7d9b55d701e754 Sep 29 17:24:13 crc kubenswrapper[4667]: I0929 17:24:13.801548 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w"] Sep 29 17:24:14 crc kubenswrapper[4667]: I0929 17:24:14.144815 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" event={"ID":"be8fb545-6b61-4464-b940-0fab07461beb","Type":"ContainerStarted","Data":"c2d162adafad4aedcd23f0f9f6c222e03ef9b5e7a202b40b6c7d9b55d701e754"} Sep 29 17:24:14 crc kubenswrapper[4667]: I0929 17:24:14.146352 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" event={"ID":"18dfc1a3-add5-49cb-860c-5361e887a24d","Type":"ContainerStarted","Data":"f3393cf0c83621bb8871f592b505c4ec3b21370ff2c0221e33589f22eafc9b27"} Sep 29 17:24:14 crc kubenswrapper[4667]: I0929 17:24:14.147893 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-66c4c8d9c4-x45lt" event={"ID":"2791e00e-ed2b-4c22-9cc1-43b616beba71","Type":"ContainerStarted","Data":"b127a67129cdc7639f02f2ca0eba6cde651a565be749e803fc89667272244674"} Sep 29 17:24:14 crc kubenswrapper[4667]: I0929 17:24:14.147952 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-66c4c8d9c4-x45lt" event={"ID":"2791e00e-ed2b-4c22-9cc1-43b616beba71","Type":"ContainerStarted","Data":"4b6f04e116922ef78313984ed3a1575bc41f4a3efc7465ab423ba786f4716783"} Sep 29 17:24:14 crc kubenswrapper[4667]: I0929 17:24:14.149199 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg" event={"ID":"b76f4f28-f053-48ff-bb61-524505148d8d","Type":"ContainerStarted","Data":"5dd6c25b2123c38f6fbb7df5da61b018357d4dcf2ae6abaa360cd39299eacd1e"} Sep 29 17:24:14 crc kubenswrapper[4667]: I0929 17:24:14.163984 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-66c4c8d9c4-x45lt" podStartSLOduration=2.16397129 podStartE2EDuration="2.16397129s" podCreationTimestamp="2025-09-29 17:24:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:24:14.158549748 +0000 UTC m=+882.656396516" watchObservedRunningTime="2025-09-29 17:24:14.16397129 +0000 UTC m=+882.661818058" Sep 29 17:24:17 crc kubenswrapper[4667]: I0929 17:24:17.167188 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hcvkt" event={"ID":"58bb8f72-34d0-44a3-a094-dc59c23fe49b","Type":"ContainerStarted","Data":"5c9b9d514948ec39b019e749137f1dcbb16bccbf16de08b954b0f69bd6093aac"} Sep 29 17:24:17 crc kubenswrapper[4667]: I0929 17:24:17.167571 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:17 crc kubenswrapper[4667]: I0929 17:24:17.168575 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg" event={"ID":"b76f4f28-f053-48ff-bb61-524505148d8d","Type":"ContainerStarted","Data":"47066e8b53e05a1265618f7794c8cb14403c787f139ef3031fb5e0eda531c502"} Sep 29 17:24:17 crc kubenswrapper[4667]: I0929 17:24:17.170115 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" event={"ID":"be8fb545-6b61-4464-b940-0fab07461beb","Type":"ContainerStarted","Data":"11dab4f3eb6513289a7e54b242f2f5a59c87cf96d6f282ff15b539f1de60a7c2"} Sep 29 17:24:17 crc kubenswrapper[4667]: I0929 17:24:17.172051 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" event={"ID":"18dfc1a3-add5-49cb-860c-5361e887a24d","Type":"ContainerStarted","Data":"7b317ce7afa6a2613cc258a19ed475e5450e570deca47514ae6e383f88096efa"} Sep 29 17:24:17 crc kubenswrapper[4667]: I0929 17:24:17.172161 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:17 crc kubenswrapper[4667]: I0929 17:24:17.181341 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-hcvkt" podStartSLOduration=1.622583997 podStartE2EDuration="5.181326062s" podCreationTimestamp="2025-09-29 17:24:12 +0000 UTC" firstStartedPulling="2025-09-29 17:24:12.879024677 +0000 UTC m=+881.376871447" lastFinishedPulling="2025-09-29 17:24:16.437766743 +0000 UTC m=+884.935613512" observedRunningTime="2025-09-29 17:24:17.177599767 +0000 UTC m=+885.675446536" watchObservedRunningTime="2025-09-29 17:24:17.181326062 +0000 UTC m=+885.679172831" Sep 29 17:24:17 crc kubenswrapper[4667]: I0929 17:24:17.188776 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-864bb6dfb5-wzvhh" podStartSLOduration=2.568851885 podStartE2EDuration="5.188762483s" podCreationTimestamp="2025-09-29 17:24:12 +0000 UTC" firstStartedPulling="2025-09-29 17:24:13.804365718 +0000 UTC m=+882.302212488" lastFinishedPulling="2025-09-29 17:24:16.424276318 +0000 UTC m=+884.922123086" observedRunningTime="2025-09-29 17:24:17.187061426 +0000 UTC m=+885.684908195" watchObservedRunningTime="2025-09-29 17:24:17.188762483 +0000 UTC m=+885.686609252" Sep 29 17:24:17 crc kubenswrapper[4667]: I0929 17:24:17.203494 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" podStartSLOduration=2.59196868 podStartE2EDuration="5.203480694s" podCreationTimestamp="2025-09-29 17:24:12 +0000 UTC" firstStartedPulling="2025-09-29 17:24:13.812590977 +0000 UTC m=+882.310437745" lastFinishedPulling="2025-09-29 17:24:16.424102991 +0000 UTC m=+884.921949759" observedRunningTime="2025-09-29 17:24:17.200689612 +0000 UTC m=+885.698536380" watchObservedRunningTime="2025-09-29 17:24:17.203480694 +0000 UTC m=+885.701327463" Sep 29 17:24:20 crc kubenswrapper[4667]: I0929 17:24:20.191597 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg" event={"ID":"b76f4f28-f053-48ff-bb61-524505148d8d","Type":"ContainerStarted","Data":"3d4f5993034df8dff2013ba66374499cc030df04dd4d165167ca574a24602282"} Sep 29 17:24:20 crc kubenswrapper[4667]: I0929 17:24:20.204258 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-58fcddf996-nvhmg" podStartSLOduration=2.134951424 podStartE2EDuration="8.204244857s" podCreationTimestamp="2025-09-29 17:24:12 +0000 UTC" firstStartedPulling="2025-09-29 17:24:13.228836673 +0000 UTC m=+881.726683442" lastFinishedPulling="2025-09-29 17:24:19.298130106 +0000 UTC m=+887.795976875" observedRunningTime="2025-09-29 17:24:20.202189141 +0000 UTC m=+888.700035910" watchObservedRunningTime="2025-09-29 17:24:20.204244857 +0000 UTC m=+888.702091626" Sep 29 17:24:22 crc kubenswrapper[4667]: I0929 17:24:22.860475 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-hcvkt" Sep 29 17:24:23 crc kubenswrapper[4667]: I0929 17:24:23.108724 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:23 crc kubenswrapper[4667]: I0929 17:24:23.108762 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:23 crc kubenswrapper[4667]: I0929 17:24:23.112785 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:23 crc kubenswrapper[4667]: I0929 17:24:23.216130 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-66c4c8d9c4-x45lt" Sep 29 17:24:23 crc kubenswrapper[4667]: I0929 17:24:23.248550 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-785bd4864d-4f7hk"] Sep 29 17:24:33 crc kubenswrapper[4667]: I0929 17:24:33.440083 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6d689559c5-h5d7w" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.413603 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4"] Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.415279 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.417008 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.424799 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4"] Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.522440 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92ffd\" (UniqueName: \"kubernetes.io/projected/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-kube-api-access-92ffd\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.522606 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.522835 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.624023 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.624099 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.624134 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92ffd\" (UniqueName: \"kubernetes.io/projected/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-kube-api-access-92ffd\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.624467 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-util\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.624522 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-bundle\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.638282 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92ffd\" (UniqueName: \"kubernetes.io/projected/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-kube-api-access-92ffd\") pod \"f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:45 crc kubenswrapper[4667]: I0929 17:24:45.727766 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:46 crc kubenswrapper[4667]: I0929 17:24:46.070981 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4"] Sep 29 17:24:46 crc kubenswrapper[4667]: I0929 17:24:46.333577 4667 generic.go:334] "Generic (PLEG): container finished" podID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerID="982734bb8fcbfe9930e55e200950691e94e100f0aa385175ecf26df8920c064d" exitCode=0 Sep 29 17:24:46 crc kubenswrapper[4667]: I0929 17:24:46.333615 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" event={"ID":"5a3413fd-e8f4-40d4-8919-87e18ffeef7d","Type":"ContainerDied","Data":"982734bb8fcbfe9930e55e200950691e94e100f0aa385175ecf26df8920c064d"} Sep 29 17:24:46 crc kubenswrapper[4667]: I0929 17:24:46.333652 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" event={"ID":"5a3413fd-e8f4-40d4-8919-87e18ffeef7d","Type":"ContainerStarted","Data":"e633f806ed86f2b7842b18fd5e878dab6cb6cfddce73e0ada268bdf826e30d8e"} Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.279465 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-785bd4864d-4f7hk" podUID="8b280c9f-d84d-4cda-a63a-800f116d0cca" containerName="console" containerID="cri-o://e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538" gracePeriod=15 Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.344999 4667 generic.go:334] "Generic (PLEG): container finished" podID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerID="1264b31f62826888f4c731a3622a6861bbfe29f836a8760982f1184636c41969" exitCode=0 Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.345085 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" event={"ID":"5a3413fd-e8f4-40d4-8919-87e18ffeef7d","Type":"ContainerDied","Data":"1264b31f62826888f4c731a3622a6861bbfe29f836a8760982f1184636c41969"} Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.622621 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-785bd4864d-4f7hk_8b280c9f-d84d-4cda-a63a-800f116d0cca/console/0.log" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.622835 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658145 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-trusted-ca-bundle\") pod \"8b280c9f-d84d-4cda-a63a-800f116d0cca\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658202 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nthfk\" (UniqueName: \"kubernetes.io/projected/8b280c9f-d84d-4cda-a63a-800f116d0cca-kube-api-access-nthfk\") pod \"8b280c9f-d84d-4cda-a63a-800f116d0cca\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658219 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-oauth-config\") pod \"8b280c9f-d84d-4cda-a63a-800f116d0cca\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658378 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-config\") pod \"8b280c9f-d84d-4cda-a63a-800f116d0cca\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658416 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-service-ca\") pod \"8b280c9f-d84d-4cda-a63a-800f116d0cca\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658510 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-serving-cert\") pod \"8b280c9f-d84d-4cda-a63a-800f116d0cca\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658568 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-oauth-serving-cert\") pod \"8b280c9f-d84d-4cda-a63a-800f116d0cca\" (UID: \"8b280c9f-d84d-4cda-a63a-800f116d0cca\") " Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658896 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "8b280c9f-d84d-4cda-a63a-800f116d0cca" (UID: "8b280c9f-d84d-4cda-a63a-800f116d0cca"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658911 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-config" (OuterVolumeSpecName: "console-config") pod "8b280c9f-d84d-4cda-a63a-800f116d0cca" (UID: "8b280c9f-d84d-4cda-a63a-800f116d0cca"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.658956 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-service-ca" (OuterVolumeSpecName: "service-ca") pod "8b280c9f-d84d-4cda-a63a-800f116d0cca" (UID: "8b280c9f-d84d-4cda-a63a-800f116d0cca"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.659180 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "8b280c9f-d84d-4cda-a63a-800f116d0cca" (UID: "8b280c9f-d84d-4cda-a63a-800f116d0cca"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.659424 4667 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.659444 4667 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-service-ca\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.659455 4667 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.659463 4667 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8b280c9f-d84d-4cda-a63a-800f116d0cca-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.670254 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b280c9f-d84d-4cda-a63a-800f116d0cca-kube-api-access-nthfk" (OuterVolumeSpecName: "kube-api-access-nthfk") pod "8b280c9f-d84d-4cda-a63a-800f116d0cca" (UID: "8b280c9f-d84d-4cda-a63a-800f116d0cca"). InnerVolumeSpecName "kube-api-access-nthfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.670290 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "8b280c9f-d84d-4cda-a63a-800f116d0cca" (UID: "8b280c9f-d84d-4cda-a63a-800f116d0cca"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.670324 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "8b280c9f-d84d-4cda-a63a-800f116d0cca" (UID: "8b280c9f-d84d-4cda-a63a-800f116d0cca"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.761631 4667 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-oauth-config\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.761668 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nthfk\" (UniqueName: \"kubernetes.io/projected/8b280c9f-d84d-4cda-a63a-800f116d0cca-kube-api-access-nthfk\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:48 crc kubenswrapper[4667]: I0929 17:24:48.761680 4667 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b280c9f-d84d-4cda-a63a-800f116d0cca-console-serving-cert\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.351688 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-785bd4864d-4f7hk_8b280c9f-d84d-4cda-a63a-800f116d0cca/console/0.log" Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.351870 4667 generic.go:334] "Generic (PLEG): container finished" podID="8b280c9f-d84d-4cda-a63a-800f116d0cca" containerID="e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538" exitCode=2 Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.351950 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-785bd4864d-4f7hk" Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.351949 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-785bd4864d-4f7hk" event={"ID":"8b280c9f-d84d-4cda-a63a-800f116d0cca","Type":"ContainerDied","Data":"e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538"} Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.352010 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-785bd4864d-4f7hk" event={"ID":"8b280c9f-d84d-4cda-a63a-800f116d0cca","Type":"ContainerDied","Data":"6e66a8c18dda8e509f1b6839f16203cb0c97886f4e4bd923f842ea17b2f8446c"} Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.352029 4667 scope.go:117] "RemoveContainer" containerID="e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538" Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.354434 4667 generic.go:334] "Generic (PLEG): container finished" podID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerID="e4d5596c84c9064693ea3d3c0aece8939be4079f95d534f56d3bd9f614172039" exitCode=0 Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.354470 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" event={"ID":"5a3413fd-e8f4-40d4-8919-87e18ffeef7d","Type":"ContainerDied","Data":"e4d5596c84c9064693ea3d3c0aece8939be4079f95d534f56d3bd9f614172039"} Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.368875 4667 scope.go:117] "RemoveContainer" containerID="e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538" Sep 29 17:24:49 crc kubenswrapper[4667]: E0929 17:24:49.369149 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538\": container with ID starting with e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538 not found: ID does not exist" containerID="e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538" Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.369211 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538"} err="failed to get container status \"e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538\": rpc error: code = NotFound desc = could not find container \"e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538\": container with ID starting with e20fa0ac72995bdbb623f1195eac31557ee57ca5bf2db6c6521e0d4da3f53538 not found: ID does not exist" Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.375573 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-785bd4864d-4f7hk"] Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.379426 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-785bd4864d-4f7hk"] Sep 29 17:24:49 crc kubenswrapper[4667]: I0929 17:24:49.822676 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b280c9f-d84d-4cda-a63a-800f116d0cca" path="/var/lib/kubelet/pods/8b280c9f-d84d-4cda-a63a-800f116d0cca/volumes" Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.580352 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.587860 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-util\") pod \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.587931 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92ffd\" (UniqueName: \"kubernetes.io/projected/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-kube-api-access-92ffd\") pod \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.587952 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-bundle\") pod \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\" (UID: \"5a3413fd-e8f4-40d4-8919-87e18ffeef7d\") " Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.588893 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-bundle" (OuterVolumeSpecName: "bundle") pod "5a3413fd-e8f4-40d4-8919-87e18ffeef7d" (UID: "5a3413fd-e8f4-40d4-8919-87e18ffeef7d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.593924 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-kube-api-access-92ffd" (OuterVolumeSpecName: "kube-api-access-92ffd") pod "5a3413fd-e8f4-40d4-8919-87e18ffeef7d" (UID: "5a3413fd-e8f4-40d4-8919-87e18ffeef7d"). InnerVolumeSpecName "kube-api-access-92ffd". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.603893 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-util" (OuterVolumeSpecName: "util") pod "5a3413fd-e8f4-40d4-8919-87e18ffeef7d" (UID: "5a3413fd-e8f4-40d4-8919-87e18ffeef7d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.689384 4667 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-util\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.689602 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92ffd\" (UniqueName: \"kubernetes.io/projected/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-kube-api-access-92ffd\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:50 crc kubenswrapper[4667]: I0929 17:24:50.689613 4667 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5a3413fd-e8f4-40d4-8919-87e18ffeef7d-bundle\") on node \"crc\" DevicePath \"\"" Sep 29 17:24:51 crc kubenswrapper[4667]: I0929 17:24:51.368742 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" event={"ID":"5a3413fd-e8f4-40d4-8919-87e18ffeef7d","Type":"ContainerDied","Data":"e633f806ed86f2b7842b18fd5e878dab6cb6cfddce73e0ada268bdf826e30d8e"} Sep 29 17:24:51 crc kubenswrapper[4667]: I0929 17:24:51.368998 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e633f806ed86f2b7842b18fd5e878dab6cb6cfddce73e0ada268bdf826e30d8e" Sep 29 17:24:51 crc kubenswrapper[4667]: I0929 17:24:51.368786 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.299683 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc"] Sep 29 17:24:59 crc kubenswrapper[4667]: E0929 17:24:59.300254 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerName="pull" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.300267 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerName="pull" Sep 29 17:24:59 crc kubenswrapper[4667]: E0929 17:24:59.300284 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b280c9f-d84d-4cda-a63a-800f116d0cca" containerName="console" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.300289 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b280c9f-d84d-4cda-a63a-800f116d0cca" containerName="console" Sep 29 17:24:59 crc kubenswrapper[4667]: E0929 17:24:59.300300 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerName="util" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.300305 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerName="util" Sep 29 17:24:59 crc kubenswrapper[4667]: E0929 17:24:59.300317 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerName="extract" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.300322 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerName="extract" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.300435 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a3413fd-e8f4-40d4-8919-87e18ffeef7d" containerName="extract" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.300442 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b280c9f-d84d-4cda-a63a-800f116d0cca" containerName="console" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.303285 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.305273 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.305418 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-44wkp" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.305309 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.305514 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.306082 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.314308 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc"] Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.417769 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f632e0ed-f363-40ef-a50e-ba37e1fad5c6-webhook-cert\") pod \"metallb-operator-controller-manager-67478f5d6-55rdc\" (UID: \"f632e0ed-f363-40ef-a50e-ba37e1fad5c6\") " pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.417855 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f632e0ed-f363-40ef-a50e-ba37e1fad5c6-apiservice-cert\") pod \"metallb-operator-controller-manager-67478f5d6-55rdc\" (UID: \"f632e0ed-f363-40ef-a50e-ba37e1fad5c6\") " pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.418012 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2zrf\" (UniqueName: \"kubernetes.io/projected/f632e0ed-f363-40ef-a50e-ba37e1fad5c6-kube-api-access-m2zrf\") pod \"metallb-operator-controller-manager-67478f5d6-55rdc\" (UID: \"f632e0ed-f363-40ef-a50e-ba37e1fad5c6\") " pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.520072 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2zrf\" (UniqueName: \"kubernetes.io/projected/f632e0ed-f363-40ef-a50e-ba37e1fad5c6-kube-api-access-m2zrf\") pod \"metallb-operator-controller-manager-67478f5d6-55rdc\" (UID: \"f632e0ed-f363-40ef-a50e-ba37e1fad5c6\") " pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.520166 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f632e0ed-f363-40ef-a50e-ba37e1fad5c6-webhook-cert\") pod \"metallb-operator-controller-manager-67478f5d6-55rdc\" (UID: \"f632e0ed-f363-40ef-a50e-ba37e1fad5c6\") " pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.520196 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f632e0ed-f363-40ef-a50e-ba37e1fad5c6-apiservice-cert\") pod \"metallb-operator-controller-manager-67478f5d6-55rdc\" (UID: \"f632e0ed-f363-40ef-a50e-ba37e1fad5c6\") " pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.527624 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f632e0ed-f363-40ef-a50e-ba37e1fad5c6-apiservice-cert\") pod \"metallb-operator-controller-manager-67478f5d6-55rdc\" (UID: \"f632e0ed-f363-40ef-a50e-ba37e1fad5c6\") " pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.532072 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m"] Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.532927 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.537342 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2zrf\" (UniqueName: \"kubernetes.io/projected/f632e0ed-f363-40ef-a50e-ba37e1fad5c6-kube-api-access-m2zrf\") pod \"metallb-operator-controller-manager-67478f5d6-55rdc\" (UID: \"f632e0ed-f363-40ef-a50e-ba37e1fad5c6\") " pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.537652 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-kvbr2" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.537829 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.537978 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.539981 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m"] Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.541588 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f632e0ed-f363-40ef-a50e-ba37e1fad5c6-webhook-cert\") pod \"metallb-operator-controller-manager-67478f5d6-55rdc\" (UID: \"f632e0ed-f363-40ef-a50e-ba37e1fad5c6\") " pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.616105 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.621514 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtkxd\" (UniqueName: \"kubernetes.io/projected/22444915-afb5-4f7b-92e4-b33472abb68e-kube-api-access-dtkxd\") pod \"metallb-operator-webhook-server-84bd8cf5c8-df27m\" (UID: \"22444915-afb5-4f7b-92e4-b33472abb68e\") " pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.621672 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/22444915-afb5-4f7b-92e4-b33472abb68e-webhook-cert\") pod \"metallb-operator-webhook-server-84bd8cf5c8-df27m\" (UID: \"22444915-afb5-4f7b-92e4-b33472abb68e\") " pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.621735 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/22444915-afb5-4f7b-92e4-b33472abb68e-apiservice-cert\") pod \"metallb-operator-webhook-server-84bd8cf5c8-df27m\" (UID: \"22444915-afb5-4f7b-92e4-b33472abb68e\") " pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.723839 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/22444915-afb5-4f7b-92e4-b33472abb68e-webhook-cert\") pod \"metallb-operator-webhook-server-84bd8cf5c8-df27m\" (UID: \"22444915-afb5-4f7b-92e4-b33472abb68e\") " pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.724485 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/22444915-afb5-4f7b-92e4-b33472abb68e-apiservice-cert\") pod \"metallb-operator-webhook-server-84bd8cf5c8-df27m\" (UID: \"22444915-afb5-4f7b-92e4-b33472abb68e\") " pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.724786 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtkxd\" (UniqueName: \"kubernetes.io/projected/22444915-afb5-4f7b-92e4-b33472abb68e-kube-api-access-dtkxd\") pod \"metallb-operator-webhook-server-84bd8cf5c8-df27m\" (UID: \"22444915-afb5-4f7b-92e4-b33472abb68e\") " pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.745817 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/22444915-afb5-4f7b-92e4-b33472abb68e-webhook-cert\") pod \"metallb-operator-webhook-server-84bd8cf5c8-df27m\" (UID: \"22444915-afb5-4f7b-92e4-b33472abb68e\") " pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.746268 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/22444915-afb5-4f7b-92e4-b33472abb68e-apiservice-cert\") pod \"metallb-operator-webhook-server-84bd8cf5c8-df27m\" (UID: \"22444915-afb5-4f7b-92e4-b33472abb68e\") " pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.765684 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtkxd\" (UniqueName: \"kubernetes.io/projected/22444915-afb5-4f7b-92e4-b33472abb68e-kube-api-access-dtkxd\") pod \"metallb-operator-webhook-server-84bd8cf5c8-df27m\" (UID: \"22444915-afb5-4f7b-92e4-b33472abb68e\") " pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:24:59 crc kubenswrapper[4667]: I0929 17:24:59.897538 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:25:00 crc kubenswrapper[4667]: I0929 17:25:00.113821 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc"] Sep 29 17:25:00 crc kubenswrapper[4667]: I0929 17:25:00.291135 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m"] Sep 29 17:25:00 crc kubenswrapper[4667]: W0929 17:25:00.301013 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22444915_afb5_4f7b_92e4_b33472abb68e.slice/crio-1bf99619f6df3cacf6a97b54620f7cd9a1b8f51d755771d9418cf3b6a49c4e7e WatchSource:0}: Error finding container 1bf99619f6df3cacf6a97b54620f7cd9a1b8f51d755771d9418cf3b6a49c4e7e: Status 404 returned error can't find the container with id 1bf99619f6df3cacf6a97b54620f7cd9a1b8f51d755771d9418cf3b6a49c4e7e Sep 29 17:25:00 crc kubenswrapper[4667]: I0929 17:25:00.421377 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" event={"ID":"22444915-afb5-4f7b-92e4-b33472abb68e","Type":"ContainerStarted","Data":"1bf99619f6df3cacf6a97b54620f7cd9a1b8f51d755771d9418cf3b6a49c4e7e"} Sep 29 17:25:00 crc kubenswrapper[4667]: I0929 17:25:00.423455 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" event={"ID":"f632e0ed-f363-40ef-a50e-ba37e1fad5c6","Type":"ContainerStarted","Data":"91c08f122fdef039db3ca1c00c329df4d4122d5100ff3c08a30510919a12972b"} Sep 29 17:25:05 crc kubenswrapper[4667]: I0929 17:25:05.485492 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" event={"ID":"22444915-afb5-4f7b-92e4-b33472abb68e","Type":"ContainerStarted","Data":"87e1082ae46d69cd623db00f989f9d329b5e2cb241032d664efe3f195b02122f"} Sep 29 17:25:05 crc kubenswrapper[4667]: I0929 17:25:05.486074 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:25:05 crc kubenswrapper[4667]: I0929 17:25:05.488448 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" event={"ID":"f632e0ed-f363-40ef-a50e-ba37e1fad5c6","Type":"ContainerStarted","Data":"72395c0f2689dec66e5e3f491f2adf5acb2c73a131cde1a8de9ed90b7636f9ff"} Sep 29 17:25:05 crc kubenswrapper[4667]: I0929 17:25:05.488580 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:25:05 crc kubenswrapper[4667]: I0929 17:25:05.507985 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" podStartSLOduration=2.437760896 podStartE2EDuration="6.5079749s" podCreationTimestamp="2025-09-29 17:24:59 +0000 UTC" firstStartedPulling="2025-09-29 17:25:00.304237588 +0000 UTC m=+928.802084356" lastFinishedPulling="2025-09-29 17:25:04.37445159 +0000 UTC m=+932.872298360" observedRunningTime="2025-09-29 17:25:05.504822428 +0000 UTC m=+934.002669196" watchObservedRunningTime="2025-09-29 17:25:05.5079749 +0000 UTC m=+934.005821670" Sep 29 17:25:05 crc kubenswrapper[4667]: I0929 17:25:05.526306 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" podStartSLOduration=2.297744991 podStartE2EDuration="6.526295233s" podCreationTimestamp="2025-09-29 17:24:59 +0000 UTC" firstStartedPulling="2025-09-29 17:25:00.124837696 +0000 UTC m=+928.622684464" lastFinishedPulling="2025-09-29 17:25:04.353387937 +0000 UTC m=+932.851234706" observedRunningTime="2025-09-29 17:25:05.523072557 +0000 UTC m=+934.020919325" watchObservedRunningTime="2025-09-29 17:25:05.526295233 +0000 UTC m=+934.024142001" Sep 29 17:25:19 crc kubenswrapper[4667]: I0929 17:25:19.901871 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-84bd8cf5c8-df27m" Sep 29 17:25:39 crc kubenswrapper[4667]: I0929 17:25:39.618812 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-67478f5d6-55rdc" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.164564 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-tbwpr"] Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.169511 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.175012 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.175194 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-28thg" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.175319 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.180609 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9"] Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.181503 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.182794 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.200310 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9"] Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.250749 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-wwv5m"] Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.250998 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qht4\" (UniqueName: \"kubernetes.io/projected/3e4ad89e-146c-4c00-b666-1ebf9487ef3a-kube-api-access-8qht4\") pod \"frr-k8s-webhook-server-5478bdb765-xmrg9\" (UID: \"3e4ad89e-146c-4c00-b666-1ebf9487ef3a\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.251054 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-frr-conf\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.251092 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-frr-sockets\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.251108 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d41412a0-9af7-438a-b70d-a391bd35670d-frr-startup\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.251134 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-reloader\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.251155 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d41412a0-9af7-438a-b70d-a391bd35670d-metrics-certs\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.251196 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e4ad89e-146c-4c00-b666-1ebf9487ef3a-cert\") pod \"frr-k8s-webhook-server-5478bdb765-xmrg9\" (UID: \"3e4ad89e-146c-4c00-b666-1ebf9487ef3a\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.251212 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7nb5\" (UniqueName: \"kubernetes.io/projected/d41412a0-9af7-438a-b70d-a391bd35670d-kube-api-access-x7nb5\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.251236 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-metrics\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.251908 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.256113 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.256272 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.256388 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5gbc9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.258307 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.259541 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5d688f5ffc-gbfh9"] Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.260518 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.263967 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.271391 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-gbfh9"] Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352373 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccf75\" (UniqueName: \"kubernetes.io/projected/c1afbef4-4d2f-4536-a4bc-5516962e56be-kube-api-access-ccf75\") pod \"controller-5d688f5ffc-gbfh9\" (UID: \"c1afbef4-4d2f-4536-a4bc-5516962e56be\") " pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352420 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qht4\" (UniqueName: \"kubernetes.io/projected/3e4ad89e-146c-4c00-b666-1ebf9487ef3a-kube-api-access-8qht4\") pod \"frr-k8s-webhook-server-5478bdb765-xmrg9\" (UID: \"3e4ad89e-146c-4c00-b666-1ebf9487ef3a\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352449 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1afbef4-4d2f-4536-a4bc-5516962e56be-cert\") pod \"controller-5d688f5ffc-gbfh9\" (UID: \"c1afbef4-4d2f-4536-a4bc-5516962e56be\") " pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352478 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-frr-conf\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352506 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-metrics-certs\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352525 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-frr-sockets\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352539 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d41412a0-9af7-438a-b70d-a391bd35670d-frr-startup\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352556 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1afbef4-4d2f-4536-a4bc-5516962e56be-metrics-certs\") pod \"controller-5d688f5ffc-gbfh9\" (UID: \"c1afbef4-4d2f-4536-a4bc-5516962e56be\") " pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352582 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-reloader\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352604 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d41412a0-9af7-438a-b70d-a391bd35670d-metrics-certs\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352621 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-metallb-excludel2\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352639 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dwxq\" (UniqueName: \"kubernetes.io/projected/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-kube-api-access-7dwxq\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352680 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e4ad89e-146c-4c00-b666-1ebf9487ef3a-cert\") pod \"frr-k8s-webhook-server-5478bdb765-xmrg9\" (UID: \"3e4ad89e-146c-4c00-b666-1ebf9487ef3a\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352698 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7nb5\" (UniqueName: \"kubernetes.io/projected/d41412a0-9af7-438a-b70d-a391bd35670d-kube-api-access-x7nb5\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352719 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-memberlist\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.352732 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-metrics\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.353116 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-metrics\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.353520 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-frr-conf\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.353705 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-frr-sockets\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.354383 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d41412a0-9af7-438a-b70d-a391bd35670d-frr-startup\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.354564 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d41412a0-9af7-438a-b70d-a391bd35670d-reloader\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: E0929 17:25:40.355101 4667 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Sep 29 17:25:40 crc kubenswrapper[4667]: E0929 17:25:40.355180 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e4ad89e-146c-4c00-b666-1ebf9487ef3a-cert podName:3e4ad89e-146c-4c00-b666-1ebf9487ef3a nodeName:}" failed. No retries permitted until 2025-09-29 17:25:40.855162622 +0000 UTC m=+969.353009391 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3e4ad89e-146c-4c00-b666-1ebf9487ef3a-cert") pod "frr-k8s-webhook-server-5478bdb765-xmrg9" (UID: "3e4ad89e-146c-4c00-b666-1ebf9487ef3a") : secret "frr-k8s-webhook-server-cert" not found Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.359415 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d41412a0-9af7-438a-b70d-a391bd35670d-metrics-certs\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.369242 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7nb5\" (UniqueName: \"kubernetes.io/projected/d41412a0-9af7-438a-b70d-a391bd35670d-kube-api-access-x7nb5\") pod \"frr-k8s-tbwpr\" (UID: \"d41412a0-9af7-438a-b70d-a391bd35670d\") " pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.374449 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qht4\" (UniqueName: \"kubernetes.io/projected/3e4ad89e-146c-4c00-b666-1ebf9487ef3a-kube-api-access-8qht4\") pod \"frr-k8s-webhook-server-5478bdb765-xmrg9\" (UID: \"3e4ad89e-146c-4c00-b666-1ebf9487ef3a\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.453415 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccf75\" (UniqueName: \"kubernetes.io/projected/c1afbef4-4d2f-4536-a4bc-5516962e56be-kube-api-access-ccf75\") pod \"controller-5d688f5ffc-gbfh9\" (UID: \"c1afbef4-4d2f-4536-a4bc-5516962e56be\") " pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.453465 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1afbef4-4d2f-4536-a4bc-5516962e56be-cert\") pod \"controller-5d688f5ffc-gbfh9\" (UID: \"c1afbef4-4d2f-4536-a4bc-5516962e56be\") " pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.453502 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-metrics-certs\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.453524 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1afbef4-4d2f-4536-a4bc-5516962e56be-metrics-certs\") pod \"controller-5d688f5ffc-gbfh9\" (UID: \"c1afbef4-4d2f-4536-a4bc-5516962e56be\") " pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.453559 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-metallb-excludel2\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.453576 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dwxq\" (UniqueName: \"kubernetes.io/projected/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-kube-api-access-7dwxq\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.453620 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-memberlist\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: E0929 17:25:40.453733 4667 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 29 17:25:40 crc kubenswrapper[4667]: E0929 17:25:40.453771 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-memberlist podName:8d9654e1-f7cf-4e80-95a8-9a133ca2e231 nodeName:}" failed. No retries permitted until 2025-09-29 17:25:40.953759993 +0000 UTC m=+969.451606762 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-memberlist") pod "speaker-wwv5m" (UID: "8d9654e1-f7cf-4e80-95a8-9a133ca2e231") : secret "metallb-memberlist" not found Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.455086 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-metallb-excludel2\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.456168 4667 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.456525 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-metrics-certs\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.461391 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1afbef4-4d2f-4536-a4bc-5516962e56be-metrics-certs\") pod \"controller-5d688f5ffc-gbfh9\" (UID: \"c1afbef4-4d2f-4536-a4bc-5516962e56be\") " pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.466617 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1afbef4-4d2f-4536-a4bc-5516962e56be-cert\") pod \"controller-5d688f5ffc-gbfh9\" (UID: \"c1afbef4-4d2f-4536-a4bc-5516962e56be\") " pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.469049 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dwxq\" (UniqueName: \"kubernetes.io/projected/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-kube-api-access-7dwxq\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.474595 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccf75\" (UniqueName: \"kubernetes.io/projected/c1afbef4-4d2f-4536-a4bc-5516962e56be-kube-api-access-ccf75\") pod \"controller-5d688f5ffc-gbfh9\" (UID: \"c1afbef4-4d2f-4536-a4bc-5516962e56be\") " pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.487173 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.580017 4667 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.595521 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.682398 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerStarted","Data":"840b0112698d4db12ae3f108f1d1571f1260ea59d8bb66a27d1d8be9cc6e63c1"} Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.859781 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e4ad89e-146c-4c00-b666-1ebf9487ef3a-cert\") pod \"frr-k8s-webhook-server-5478bdb765-xmrg9\" (UID: \"3e4ad89e-146c-4c00-b666-1ebf9487ef3a\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.863899 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3e4ad89e-146c-4c00-b666-1ebf9487ef3a-cert\") pod \"frr-k8s-webhook-server-5478bdb765-xmrg9\" (UID: \"3e4ad89e-146c-4c00-b666-1ebf9487ef3a\") " pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.935202 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5d688f5ffc-gbfh9"] Sep 29 17:25:40 crc kubenswrapper[4667]: W0929 17:25:40.938387 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1afbef4_4d2f_4536_a4bc_5516962e56be.slice/crio-8f43187031a76552416105b7f3dc7b80bf3b5ef5f8ddfc130faf28f7f7aa11a0 WatchSource:0}: Error finding container 8f43187031a76552416105b7f3dc7b80bf3b5ef5f8ddfc130faf28f7f7aa11a0: Status 404 returned error can't find the container with id 8f43187031a76552416105b7f3dc7b80bf3b5ef5f8ddfc130faf28f7f7aa11a0 Sep 29 17:25:40 crc kubenswrapper[4667]: I0929 17:25:40.961376 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-memberlist\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:40 crc kubenswrapper[4667]: E0929 17:25:40.961495 4667 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Sep 29 17:25:40 crc kubenswrapper[4667]: E0929 17:25:40.961543 4667 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-memberlist podName:8d9654e1-f7cf-4e80-95a8-9a133ca2e231 nodeName:}" failed. No retries permitted until 2025-09-29 17:25:41.961531835 +0000 UTC m=+970.459378604 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-memberlist") pod "speaker-wwv5m" (UID: "8d9654e1-f7cf-4e80-95a8-9a133ca2e231") : secret "metallb-memberlist" not found Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.102099 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.449618 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9"] Sep 29 17:25:41 crc kubenswrapper[4667]: W0929 17:25:41.452058 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e4ad89e_146c_4c00_b666_1ebf9487ef3a.slice/crio-17be6bfe3c4292d5e70dbd95d9ca8ae6febaa084c3d06481409f4d3efd1219fa WatchSource:0}: Error finding container 17be6bfe3c4292d5e70dbd95d9ca8ae6febaa084c3d06481409f4d3efd1219fa: Status 404 returned error can't find the container with id 17be6bfe3c4292d5e70dbd95d9ca8ae6febaa084c3d06481409f4d3efd1219fa Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.688301 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" event={"ID":"3e4ad89e-146c-4c00-b666-1ebf9487ef3a","Type":"ContainerStarted","Data":"17be6bfe3c4292d5e70dbd95d9ca8ae6febaa084c3d06481409f4d3efd1219fa"} Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.689976 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-gbfh9" event={"ID":"c1afbef4-4d2f-4536-a4bc-5516962e56be","Type":"ContainerStarted","Data":"ddf9bde3f3b004c1bee7b9733050265e1786b85830d9af6020c0a4e8f2acb094"} Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.690017 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-gbfh9" event={"ID":"c1afbef4-4d2f-4536-a4bc-5516962e56be","Type":"ContainerStarted","Data":"9cb9abe2b5708a618937f163968799c2634d27c4393035c0cf6e51f908e1a1db"} Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.690028 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5d688f5ffc-gbfh9" event={"ID":"c1afbef4-4d2f-4536-a4bc-5516962e56be","Type":"ContainerStarted","Data":"8f43187031a76552416105b7f3dc7b80bf3b5ef5f8ddfc130faf28f7f7aa11a0"} Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.690091 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.703303 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5d688f5ffc-gbfh9" podStartSLOduration=1.703290309 podStartE2EDuration="1.703290309s" podCreationTimestamp="2025-09-29 17:25:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:25:41.700567134 +0000 UTC m=+970.198413903" watchObservedRunningTime="2025-09-29 17:25:41.703290309 +0000 UTC m=+970.201137078" Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.975814 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-memberlist\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:41 crc kubenswrapper[4667]: I0929 17:25:41.979619 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8d9654e1-f7cf-4e80-95a8-9a133ca2e231-memberlist\") pod \"speaker-wwv5m\" (UID: \"8d9654e1-f7cf-4e80-95a8-9a133ca2e231\") " pod="metallb-system/speaker-wwv5m" Sep 29 17:25:42 crc kubenswrapper[4667]: I0929 17:25:42.069111 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wwv5m" Sep 29 17:25:42 crc kubenswrapper[4667]: W0929 17:25:42.086154 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d9654e1_f7cf_4e80_95a8_9a133ca2e231.slice/crio-45174793a6c2ca54ba78a9a19e31d0d99f59ee053204e13435fb7f420831d077 WatchSource:0}: Error finding container 45174793a6c2ca54ba78a9a19e31d0d99f59ee053204e13435fb7f420831d077: Status 404 returned error can't find the container with id 45174793a6c2ca54ba78a9a19e31d0d99f59ee053204e13435fb7f420831d077 Sep 29 17:25:42 crc kubenswrapper[4667]: I0929 17:25:42.706830 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wwv5m" event={"ID":"8d9654e1-f7cf-4e80-95a8-9a133ca2e231","Type":"ContainerStarted","Data":"6d8a410ef08417b6f80a773130cc53dff459eeef280e67a1a180dbbffad632d3"} Sep 29 17:25:42 crc kubenswrapper[4667]: I0929 17:25:42.706884 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wwv5m" event={"ID":"8d9654e1-f7cf-4e80-95a8-9a133ca2e231","Type":"ContainerStarted","Data":"a42a6ad2d7646aa00d07a9c2fea484a71844c840f74471989a906ceb2d5864e3"} Sep 29 17:25:42 crc kubenswrapper[4667]: I0929 17:25:42.706895 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wwv5m" event={"ID":"8d9654e1-f7cf-4e80-95a8-9a133ca2e231","Type":"ContainerStarted","Data":"45174793a6c2ca54ba78a9a19e31d0d99f59ee053204e13435fb7f420831d077"} Sep 29 17:25:42 crc kubenswrapper[4667]: I0929 17:25:42.707104 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-wwv5m" Sep 29 17:25:47 crc kubenswrapper[4667]: I0929 17:25:47.745226 4667 generic.go:334] "Generic (PLEG): container finished" podID="d41412a0-9af7-438a-b70d-a391bd35670d" containerID="cdadcf4539f24dd4af01065a05e5fe56255814ba822c27e421701ec9d54ce23f" exitCode=0 Sep 29 17:25:47 crc kubenswrapper[4667]: I0929 17:25:47.745473 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerDied","Data":"cdadcf4539f24dd4af01065a05e5fe56255814ba822c27e421701ec9d54ce23f"} Sep 29 17:25:47 crc kubenswrapper[4667]: I0929 17:25:47.748007 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" event={"ID":"3e4ad89e-146c-4c00-b666-1ebf9487ef3a","Type":"ContainerStarted","Data":"47d32487d0a03566d67fca01d8fe8c166012b2313118728a88cdd4568d1f0c5f"} Sep 29 17:25:47 crc kubenswrapper[4667]: I0929 17:25:47.748837 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:25:47 crc kubenswrapper[4667]: I0929 17:25:47.790550 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-wwv5m" podStartSLOduration=7.790529468 podStartE2EDuration="7.790529468s" podCreationTimestamp="2025-09-29 17:25:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-29 17:25:42.718322448 +0000 UTC m=+971.216169228" watchObservedRunningTime="2025-09-29 17:25:47.790529468 +0000 UTC m=+976.288376237" Sep 29 17:25:47 crc kubenswrapper[4667]: I0929 17:25:47.821344 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" podStartSLOduration=1.699395105 podStartE2EDuration="7.821324891s" podCreationTimestamp="2025-09-29 17:25:40 +0000 UTC" firstStartedPulling="2025-09-29 17:25:41.453830573 +0000 UTC m=+969.951677343" lastFinishedPulling="2025-09-29 17:25:47.575760349 +0000 UTC m=+976.073607129" observedRunningTime="2025-09-29 17:25:47.815692652 +0000 UTC m=+976.313539422" watchObservedRunningTime="2025-09-29 17:25:47.821324891 +0000 UTC m=+976.319171660" Sep 29 17:25:48 crc kubenswrapper[4667]: I0929 17:25:48.754257 4667 generic.go:334] "Generic (PLEG): container finished" podID="d41412a0-9af7-438a-b70d-a391bd35670d" containerID="7c04a892465efd839492694409958b82324b6f9604b8af33c639832901ffce38" exitCode=0 Sep 29 17:25:48 crc kubenswrapper[4667]: I0929 17:25:48.754353 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerDied","Data":"7c04a892465efd839492694409958b82324b6f9604b8af33c639832901ffce38"} Sep 29 17:25:49 crc kubenswrapper[4667]: I0929 17:25:49.763069 4667 generic.go:334] "Generic (PLEG): container finished" podID="d41412a0-9af7-438a-b70d-a391bd35670d" containerID="284929b9a8c6b4cdbe6e52e6987b668fd6c715ac936471eb6da65ca04bb71c73" exitCode=0 Sep 29 17:25:49 crc kubenswrapper[4667]: I0929 17:25:49.763142 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerDied","Data":"284929b9a8c6b4cdbe6e52e6987b668fd6c715ac936471eb6da65ca04bb71c73"} Sep 29 17:25:50 crc kubenswrapper[4667]: I0929 17:25:50.774936 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerStarted","Data":"a8c88cdd75edb0fe70d02848c3eda432445ae967083103b7726a6644ff8da3e1"} Sep 29 17:25:50 crc kubenswrapper[4667]: I0929 17:25:50.775152 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerStarted","Data":"302558ebc65223bd2a6564d001a7723a5a1257596653924fda724d6f16823099"} Sep 29 17:25:50 crc kubenswrapper[4667]: I0929 17:25:50.775170 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:50 crc kubenswrapper[4667]: I0929 17:25:50.775180 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerStarted","Data":"242931cc80dbdcb6addb5a68833272682f28ef1c5a72bc609a72f15b88f4c223"} Sep 29 17:25:50 crc kubenswrapper[4667]: I0929 17:25:50.775190 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerStarted","Data":"2dbd240b7fe3bde2d76a8a5970ed555887681cf67341ed5376923f0e62235fd6"} Sep 29 17:25:50 crc kubenswrapper[4667]: I0929 17:25:50.775197 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerStarted","Data":"6dea67cd8efeca937783f872442d883d4f98067f01ba0790339d3a47978cbf92"} Sep 29 17:25:50 crc kubenswrapper[4667]: I0929 17:25:50.775204 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-tbwpr" event={"ID":"d41412a0-9af7-438a-b70d-a391bd35670d","Type":"ContainerStarted","Data":"6186666b6bebea3bf8d791041f35a780dad20e604f6014b31db2aad40e3bae0b"} Sep 29 17:25:50 crc kubenswrapper[4667]: I0929 17:25:50.790819 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-tbwpr" podStartSLOduration=3.790581658 podStartE2EDuration="10.790810039s" podCreationTimestamp="2025-09-29 17:25:40 +0000 UTC" firstStartedPulling="2025-09-29 17:25:40.579777062 +0000 UTC m=+969.077623831" lastFinishedPulling="2025-09-29 17:25:47.580005443 +0000 UTC m=+976.077852212" observedRunningTime="2025-09-29 17:25:50.78818995 +0000 UTC m=+979.286036719" watchObservedRunningTime="2025-09-29 17:25:50.790810039 +0000 UTC m=+979.288656809" Sep 29 17:25:52 crc kubenswrapper[4667]: I0929 17:25:52.073407 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-wwv5m" Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.167019 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-l7ncc"] Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.167923 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l7ncc" Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.169083 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-44ht4" Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.169324 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.169415 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.181492 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-l7ncc"] Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.265320 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5x2p\" (UniqueName: \"kubernetes.io/projected/6c5a980f-afe5-4c15-ad9c-589cf97407eb-kube-api-access-m5x2p\") pod \"openstack-operator-index-l7ncc\" (UID: \"6c5a980f-afe5-4c15-ad9c-589cf97407eb\") " pod="openstack-operators/openstack-operator-index-l7ncc" Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.366937 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5x2p\" (UniqueName: \"kubernetes.io/projected/6c5a980f-afe5-4c15-ad9c-589cf97407eb-kube-api-access-m5x2p\") pod \"openstack-operator-index-l7ncc\" (UID: \"6c5a980f-afe5-4c15-ad9c-589cf97407eb\") " pod="openstack-operators/openstack-operator-index-l7ncc" Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.380717 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5x2p\" (UniqueName: \"kubernetes.io/projected/6c5a980f-afe5-4c15-ad9c-589cf97407eb-kube-api-access-m5x2p\") pod \"openstack-operator-index-l7ncc\" (UID: \"6c5a980f-afe5-4c15-ad9c-589cf97407eb\") " pod="openstack-operators/openstack-operator-index-l7ncc" Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.484611 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l7ncc" Sep 29 17:25:54 crc kubenswrapper[4667]: I0929 17:25:54.855567 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-l7ncc"] Sep 29 17:25:54 crc kubenswrapper[4667]: W0929 17:25:54.856859 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c5a980f_afe5_4c15_ad9c_589cf97407eb.slice/crio-9554e6e58872ff0f09c9b25c221c8d6f59504ae1ddf6ba8ce7259801dd521c52 WatchSource:0}: Error finding container 9554e6e58872ff0f09c9b25c221c8d6f59504ae1ddf6ba8ce7259801dd521c52: Status 404 returned error can't find the container with id 9554e6e58872ff0f09c9b25c221c8d6f59504ae1ddf6ba8ce7259801dd521c52 Sep 29 17:25:55 crc kubenswrapper[4667]: I0929 17:25:55.487900 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:55 crc kubenswrapper[4667]: I0929 17:25:55.520495 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:25:55 crc kubenswrapper[4667]: I0929 17:25:55.806800 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l7ncc" event={"ID":"6c5a980f-afe5-4c15-ad9c-589cf97407eb","Type":"ContainerStarted","Data":"9554e6e58872ff0f09c9b25c221c8d6f59504ae1ddf6ba8ce7259801dd521c52"} Sep 29 17:25:56 crc kubenswrapper[4667]: I0929 17:25:56.897550 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:25:56 crc kubenswrapper[4667]: I0929 17:25:56.897779 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:25:57 crc kubenswrapper[4667]: I0929 17:25:57.552376 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-l7ncc"] Sep 29 17:25:58 crc kubenswrapper[4667]: I0929 17:25:58.155755 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-6b7qn"] Sep 29 17:25:58 crc kubenswrapper[4667]: I0929 17:25:58.157095 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6b7qn" Sep 29 17:25:58 crc kubenswrapper[4667]: I0929 17:25:58.163896 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6b7qn"] Sep 29 17:25:58 crc kubenswrapper[4667]: I0929 17:25:58.320818 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmkl8\" (UniqueName: \"kubernetes.io/projected/7aa88991-5ef2-4166-8d37-8ce5fb207d26-kube-api-access-kmkl8\") pod \"openstack-operator-index-6b7qn\" (UID: \"7aa88991-5ef2-4166-8d37-8ce5fb207d26\") " pod="openstack-operators/openstack-operator-index-6b7qn" Sep 29 17:25:58 crc kubenswrapper[4667]: I0929 17:25:58.422316 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmkl8\" (UniqueName: \"kubernetes.io/projected/7aa88991-5ef2-4166-8d37-8ce5fb207d26-kube-api-access-kmkl8\") pod \"openstack-operator-index-6b7qn\" (UID: \"7aa88991-5ef2-4166-8d37-8ce5fb207d26\") " pod="openstack-operators/openstack-operator-index-6b7qn" Sep 29 17:25:58 crc kubenswrapper[4667]: I0929 17:25:58.436697 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmkl8\" (UniqueName: \"kubernetes.io/projected/7aa88991-5ef2-4166-8d37-8ce5fb207d26-kube-api-access-kmkl8\") pod \"openstack-operator-index-6b7qn\" (UID: \"7aa88991-5ef2-4166-8d37-8ce5fb207d26\") " pod="openstack-operators/openstack-operator-index-6b7qn" Sep 29 17:25:58 crc kubenswrapper[4667]: I0929 17:25:58.474673 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6b7qn" Sep 29 17:25:58 crc kubenswrapper[4667]: I0929 17:25:58.808557 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6b7qn"] Sep 29 17:25:58 crc kubenswrapper[4667]: W0929 17:25:58.812468 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7aa88991_5ef2_4166_8d37_8ce5fb207d26.slice/crio-5e2bd2eed6ddb3f6205f7d3f091b449f9d0de3fdf2de17a11f6f62b6003b74f0 WatchSource:0}: Error finding container 5e2bd2eed6ddb3f6205f7d3f091b449f9d0de3fdf2de17a11f6f62b6003b74f0: Status 404 returned error can't find the container with id 5e2bd2eed6ddb3f6205f7d3f091b449f9d0de3fdf2de17a11f6f62b6003b74f0 Sep 29 17:25:58 crc kubenswrapper[4667]: I0929 17:25:58.823232 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6b7qn" event={"ID":"7aa88991-5ef2-4166-8d37-8ce5fb207d26","Type":"ContainerStarted","Data":"5e2bd2eed6ddb3f6205f7d3f091b449f9d0de3fdf2de17a11f6f62b6003b74f0"} Sep 29 17:26:00 crc kubenswrapper[4667]: I0929 17:26:00.489519 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-tbwpr" Sep 29 17:26:00 crc kubenswrapper[4667]: I0929 17:26:00.598137 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5d688f5ffc-gbfh9" Sep 29 17:26:01 crc kubenswrapper[4667]: I0929 17:26:01.107834 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-5478bdb765-xmrg9" Sep 29 17:26:26 crc kubenswrapper[4667]: I0929 17:26:26.897391 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:26:26 crc kubenswrapper[4667]: I0929 17:26:26.897750 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:26:56 crc kubenswrapper[4667]: I0929 17:26:56.897013 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:26:56 crc kubenswrapper[4667]: I0929 17:26:56.897375 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:26:56 crc kubenswrapper[4667]: I0929 17:26:56.897417 4667 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:26:56 crc kubenswrapper[4667]: I0929 17:26:56.897822 4667 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"394308b51f762322854d3b0cdd5e2f611129dcf7d5f915f00008534ba0f925d4"} pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 17:26:56 crc kubenswrapper[4667]: I0929 17:26:56.897889 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" containerID="cri-o://394308b51f762322854d3b0cdd5e2f611129dcf7d5f915f00008534ba0f925d4" gracePeriod=600 Sep 29 17:26:57 crc kubenswrapper[4667]: I0929 17:26:57.128323 4667 generic.go:334] "Generic (PLEG): container finished" podID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerID="394308b51f762322854d3b0cdd5e2f611129dcf7d5f915f00008534ba0f925d4" exitCode=0 Sep 29 17:26:57 crc kubenswrapper[4667]: I0929 17:26:57.128381 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerDied","Data":"394308b51f762322854d3b0cdd5e2f611129dcf7d5f915f00008534ba0f925d4"} Sep 29 17:26:57 crc kubenswrapper[4667]: I0929 17:26:57.128495 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"94b33344ce4f405cbd10df51f10373f90996afec3b5c19edb1480b7527aa57d4"} Sep 29 17:26:57 crc kubenswrapper[4667]: I0929 17:26:57.128515 4667 scope.go:117] "RemoveContainer" containerID="24409b364de3d67a164134cbdd70a1ed09fbd088e081c52528c8d39cd38525a9" Sep 29 17:27:54 crc kubenswrapper[4667]: E0929 17:27:54.863774 4667 log.go:32] "PullImage from image service failed" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:27:54 crc kubenswrapper[4667]: E0929 17:27:54.864192 4667 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:27:54 crc kubenswrapper[4667]: E0929 17:27:54.864316 4667 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-m5x2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-operator-index-l7ncc_openstack-operators(6c5a980f-afe5-4c15-ad9c-589cf97407eb): ErrImagePull: rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" logger="UnhandledError" Sep 29 17:27:54 crc kubenswrapper[4667]: E0929 17:27:54.866192 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \\\"http://38.129.56.34:5001/v2/\\\": dial tcp 38.129.56.34:5001: i/o timeout\"" pod="openstack-operators/openstack-operator-index-l7ncc" podUID="6c5a980f-afe5-4c15-ad9c-589cf97407eb" Sep 29 17:27:55 crc kubenswrapper[4667]: I0929 17:27:55.724312 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l7ncc" Sep 29 17:27:55 crc kubenswrapper[4667]: I0929 17:27:55.838900 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5x2p\" (UniqueName: \"kubernetes.io/projected/6c5a980f-afe5-4c15-ad9c-589cf97407eb-kube-api-access-m5x2p\") pod \"6c5a980f-afe5-4c15-ad9c-589cf97407eb\" (UID: \"6c5a980f-afe5-4c15-ad9c-589cf97407eb\") " Sep 29 17:27:55 crc kubenswrapper[4667]: I0929 17:27:55.843693 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c5a980f-afe5-4c15-ad9c-589cf97407eb-kube-api-access-m5x2p" (OuterVolumeSpecName: "kube-api-access-m5x2p") pod "6c5a980f-afe5-4c15-ad9c-589cf97407eb" (UID: "6c5a980f-afe5-4c15-ad9c-589cf97407eb"). InnerVolumeSpecName "kube-api-access-m5x2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:27:55 crc kubenswrapper[4667]: I0929 17:27:55.940636 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5x2p\" (UniqueName: \"kubernetes.io/projected/6c5a980f-afe5-4c15-ad9c-589cf97407eb-kube-api-access-m5x2p\") on node \"crc\" DevicePath \"\"" Sep 29 17:27:56 crc kubenswrapper[4667]: I0929 17:27:56.458337 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-l7ncc" event={"ID":"6c5a980f-afe5-4c15-ad9c-589cf97407eb","Type":"ContainerDied","Data":"9554e6e58872ff0f09c9b25c221c8d6f59504ae1ddf6ba8ce7259801dd521c52"} Sep 29 17:27:56 crc kubenswrapper[4667]: I0929 17:27:56.458376 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-l7ncc" Sep 29 17:27:56 crc kubenswrapper[4667]: I0929 17:27:56.489039 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-l7ncc"] Sep 29 17:27:56 crc kubenswrapper[4667]: I0929 17:27:56.494744 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-l7ncc"] Sep 29 17:27:57 crc kubenswrapper[4667]: I0929 17:27:57.824205 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c5a980f-afe5-4c15-ad9c-589cf97407eb" path="/var/lib/kubelet/pods/6c5a980f-afe5-4c15-ad9c-589cf97407eb/volumes" Sep 29 17:27:58 crc kubenswrapper[4667]: E0929 17:27:58.817607 4667 log.go:32] "PullImage from image service failed" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:27:58 crc kubenswrapper[4667]: E0929 17:27:58.817955 4667 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:27:58 crc kubenswrapper[4667]: E0929 17:27:58.818074 4667 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmkl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-operator-index-6b7qn_openstack-operators(7aa88991-5ef2-4166-8d37-8ce5fb207d26): ErrImagePull: rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" logger="UnhandledError" Sep 29 17:27:58 crc kubenswrapper[4667]: E0929 17:27:58.819300 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \\\"http://38.129.56.34:5001/v2/\\\": dial tcp 38.129.56.34:5001: i/o timeout\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:27:59 crc kubenswrapper[4667]: E0929 17:27:59.477072 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:29:26 crc kubenswrapper[4667]: I0929 17:29:26.897587 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:29:26 crc kubenswrapper[4667]: I0929 17:29:26.897926 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:29:56 crc kubenswrapper[4667]: I0929 17:29:56.897923 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:29:56 crc kubenswrapper[4667]: I0929 17:29:56.898894 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.130404 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf"] Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.131484 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.132863 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.132940 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.138649 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf"] Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.268733 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpk78\" (UniqueName: \"kubernetes.io/projected/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-kube-api-access-kpk78\") pod \"collect-profiles-29319450-qxpbf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.268777 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-secret-volume\") pod \"collect-profiles-29319450-qxpbf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.268814 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-config-volume\") pod \"collect-profiles-29319450-qxpbf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.369432 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpk78\" (UniqueName: \"kubernetes.io/projected/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-kube-api-access-kpk78\") pod \"collect-profiles-29319450-qxpbf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.369473 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-secret-volume\") pod \"collect-profiles-29319450-qxpbf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.369532 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-config-volume\") pod \"collect-profiles-29319450-qxpbf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.370295 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-config-volume\") pod \"collect-profiles-29319450-qxpbf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.373886 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-secret-volume\") pod \"collect-profiles-29319450-qxpbf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.381899 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpk78\" (UniqueName: \"kubernetes.io/projected/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-kube-api-access-kpk78\") pod \"collect-profiles-29319450-qxpbf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.444918 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:00 crc kubenswrapper[4667]: I0929 17:30:00.794802 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf"] Sep 29 17:30:01 crc kubenswrapper[4667]: I0929 17:30:01.106657 4667 generic.go:334] "Generic (PLEG): container finished" podID="3a8bedd9-4bc1-4992-b0e0-b35725a5febf" containerID="e0d391f20a28bd440cbadee110194111ed936369cca1444cad2b57bec8138c45" exitCode=0 Sep 29 17:30:01 crc kubenswrapper[4667]: I0929 17:30:01.106692 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" event={"ID":"3a8bedd9-4bc1-4992-b0e0-b35725a5febf","Type":"ContainerDied","Data":"e0d391f20a28bd440cbadee110194111ed936369cca1444cad2b57bec8138c45"} Sep 29 17:30:01 crc kubenswrapper[4667]: I0929 17:30:01.106715 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" event={"ID":"3a8bedd9-4bc1-4992-b0e0-b35725a5febf","Type":"ContainerStarted","Data":"a80a10182702b60b99dafa2d3f09f81d29ed72e4bf0b96ed13050b32f059b142"} Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.310837 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.498287 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpk78\" (UniqueName: \"kubernetes.io/projected/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-kube-api-access-kpk78\") pod \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.498395 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-secret-volume\") pod \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.498539 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-config-volume\") pod \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\" (UID: \"3a8bedd9-4bc1-4992-b0e0-b35725a5febf\") " Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.499164 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-config-volume" (OuterVolumeSpecName: "config-volume") pod "3a8bedd9-4bc1-4992-b0e0-b35725a5febf" (UID: "3a8bedd9-4bc1-4992-b0e0-b35725a5febf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.502304 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3a8bedd9-4bc1-4992-b0e0-b35725a5febf" (UID: "3a8bedd9-4bc1-4992-b0e0-b35725a5febf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.502455 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-kube-api-access-kpk78" (OuterVolumeSpecName: "kube-api-access-kpk78") pod "3a8bedd9-4bc1-4992-b0e0-b35725a5febf" (UID: "3a8bedd9-4bc1-4992-b0e0-b35725a5febf"). InnerVolumeSpecName "kube-api-access-kpk78". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.599770 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpk78\" (UniqueName: \"kubernetes.io/projected/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-kube-api-access-kpk78\") on node \"crc\" DevicePath \"\"" Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.599794 4667 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 17:30:02 crc kubenswrapper[4667]: I0929 17:30:02.599804 4667 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3a8bedd9-4bc1-4992-b0e0-b35725a5febf-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 17:30:03 crc kubenswrapper[4667]: I0929 17:30:03.117996 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" event={"ID":"3a8bedd9-4bc1-4992-b0e0-b35725a5febf","Type":"ContainerDied","Data":"a80a10182702b60b99dafa2d3f09f81d29ed72e4bf0b96ed13050b32f059b142"} Sep 29 17:30:03 crc kubenswrapper[4667]: I0929 17:30:03.118034 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a80a10182702b60b99dafa2d3f09f81d29ed72e4bf0b96ed13050b32f059b142" Sep 29 17:30:03 crc kubenswrapper[4667]: I0929 17:30:03.118057 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319450-qxpbf" Sep 29 17:30:11 crc kubenswrapper[4667]: E0929 17:30:11.823314 4667 log.go:32] "PullImage from image service failed" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:30:11 crc kubenswrapper[4667]: E0929 17:30:11.823668 4667 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:30:11 crc kubenswrapper[4667]: E0929 17:30:11.823779 4667 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmkl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-operator-index-6b7qn_openstack-operators(7aa88991-5ef2-4166-8d37-8ce5fb207d26): ErrImagePull: rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" logger="UnhandledError" Sep 29 17:30:11 crc kubenswrapper[4667]: E0929 17:30:11.824949 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \\\"http://38.129.56.34:5001/v2/\\\": dial tcp 38.129.56.34:5001: i/o timeout\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:30:24 crc kubenswrapper[4667]: E0929 17:30:24.817309 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:30:26 crc kubenswrapper[4667]: I0929 17:30:26.897400 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:30:26 crc kubenswrapper[4667]: I0929 17:30:26.897629 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:30:26 crc kubenswrapper[4667]: I0929 17:30:26.897669 4667 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:30:26 crc kubenswrapper[4667]: I0929 17:30:26.898156 4667 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"94b33344ce4f405cbd10df51f10373f90996afec3b5c19edb1480b7527aa57d4"} pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 17:30:26 crc kubenswrapper[4667]: I0929 17:30:26.898208 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" containerID="cri-o://94b33344ce4f405cbd10df51f10373f90996afec3b5c19edb1480b7527aa57d4" gracePeriod=600 Sep 29 17:30:27 crc kubenswrapper[4667]: I0929 17:30:27.233792 4667 generic.go:334] "Generic (PLEG): container finished" podID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerID="94b33344ce4f405cbd10df51f10373f90996afec3b5c19edb1480b7527aa57d4" exitCode=0 Sep 29 17:30:27 crc kubenswrapper[4667]: I0929 17:30:27.233882 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerDied","Data":"94b33344ce4f405cbd10df51f10373f90996afec3b5c19edb1480b7527aa57d4"} Sep 29 17:30:27 crc kubenswrapper[4667]: I0929 17:30:27.234020 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"9c9d598e95ffa725e3f0ee5b1aa348c0acb3951aa004315e545eed19b8702772"} Sep 29 17:30:27 crc kubenswrapper[4667]: I0929 17:30:27.234043 4667 scope.go:117] "RemoveContainer" containerID="394308b51f762322854d3b0cdd5e2f611129dcf7d5f915f00008534ba0f925d4" Sep 29 17:32:38 crc kubenswrapper[4667]: E0929 17:32:38.823031 4667 log.go:32] "PullImage from image service failed" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:32:38 crc kubenswrapper[4667]: E0929 17:32:38.823691 4667 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:32:38 crc kubenswrapper[4667]: E0929 17:32:38.823930 4667 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmkl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-operator-index-6b7qn_openstack-operators(7aa88991-5ef2-4166-8d37-8ce5fb207d26): ErrImagePull: rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" logger="UnhandledError" Sep 29 17:32:38 crc kubenswrapper[4667]: E0929 17:32:38.825116 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \\\"http://38.129.56.34:5001/v2/\\\": dial tcp 38.129.56.34:5001: i/o timeout\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:32:53 crc kubenswrapper[4667]: E0929 17:32:53.817005 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:32:56 crc kubenswrapper[4667]: I0929 17:32:56.897709 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:32:56 crc kubenswrapper[4667]: I0929 17:32:56.898543 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:32:59 crc kubenswrapper[4667]: I0929 17:32:59.999608 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rtfrc"] Sep 29 17:33:00 crc kubenswrapper[4667]: E0929 17:33:00.000179 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8bedd9-4bc1-4992-b0e0-b35725a5febf" containerName="collect-profiles" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.000191 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8bedd9-4bc1-4992-b0e0-b35725a5febf" containerName="collect-profiles" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.000319 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a8bedd9-4bc1-4992-b0e0-b35725a5febf" containerName="collect-profiles" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.001205 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.005484 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rtfrc"] Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.163406 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg987\" (UniqueName: \"kubernetes.io/projected/9c7d713d-a85d-4a36-a338-d35fcb516b24-kube-api-access-hg987\") pod \"redhat-operators-rtfrc\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.163599 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-catalog-content\") pod \"redhat-operators-rtfrc\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.163629 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-utilities\") pod \"redhat-operators-rtfrc\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.264770 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-catalog-content\") pod \"redhat-operators-rtfrc\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.264811 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-utilities\") pod \"redhat-operators-rtfrc\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.264886 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg987\" (UniqueName: \"kubernetes.io/projected/9c7d713d-a85d-4a36-a338-d35fcb516b24-kube-api-access-hg987\") pod \"redhat-operators-rtfrc\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.265204 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-catalog-content\") pod \"redhat-operators-rtfrc\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.265242 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-utilities\") pod \"redhat-operators-rtfrc\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.280604 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg987\" (UniqueName: \"kubernetes.io/projected/9c7d713d-a85d-4a36-a338-d35fcb516b24-kube-api-access-hg987\") pod \"redhat-operators-rtfrc\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.314329 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:00 crc kubenswrapper[4667]: I0929 17:33:00.675278 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rtfrc"] Sep 29 17:33:01 crc kubenswrapper[4667]: I0929 17:33:01.062000 4667 generic.go:334] "Generic (PLEG): container finished" podID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerID="8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022" exitCode=0 Sep 29 17:33:01 crc kubenswrapper[4667]: I0929 17:33:01.062054 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtfrc" event={"ID":"9c7d713d-a85d-4a36-a338-d35fcb516b24","Type":"ContainerDied","Data":"8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022"} Sep 29 17:33:01 crc kubenswrapper[4667]: I0929 17:33:01.062166 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtfrc" event={"ID":"9c7d713d-a85d-4a36-a338-d35fcb516b24","Type":"ContainerStarted","Data":"5dc4ad4be77d1151e759971efebcad8b9d2e2ac4e0d2165283da31a40182c649"} Sep 29 17:33:01 crc kubenswrapper[4667]: I0929 17:33:01.063355 4667 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 17:33:02 crc kubenswrapper[4667]: I0929 17:33:02.068480 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtfrc" event={"ID":"9c7d713d-a85d-4a36-a338-d35fcb516b24","Type":"ContainerStarted","Data":"44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1"} Sep 29 17:33:03 crc kubenswrapper[4667]: I0929 17:33:03.076290 4667 generic.go:334] "Generic (PLEG): container finished" podID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerID="44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1" exitCode=0 Sep 29 17:33:03 crc kubenswrapper[4667]: I0929 17:33:03.076353 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtfrc" event={"ID":"9c7d713d-a85d-4a36-a338-d35fcb516b24","Type":"ContainerDied","Data":"44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1"} Sep 29 17:33:04 crc kubenswrapper[4667]: I0929 17:33:04.083417 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtfrc" event={"ID":"9c7d713d-a85d-4a36-a338-d35fcb516b24","Type":"ContainerStarted","Data":"d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61"} Sep 29 17:33:04 crc kubenswrapper[4667]: I0929 17:33:04.097322 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rtfrc" podStartSLOduration=2.492465358 podStartE2EDuration="5.097304336s" podCreationTimestamp="2025-09-29 17:32:59 +0000 UTC" firstStartedPulling="2025-09-29 17:33:01.063155994 +0000 UTC m=+1409.561002762" lastFinishedPulling="2025-09-29 17:33:03.667994971 +0000 UTC m=+1412.165841740" observedRunningTime="2025-09-29 17:33:04.094905115 +0000 UTC m=+1412.592751884" watchObservedRunningTime="2025-09-29 17:33:04.097304336 +0000 UTC m=+1412.595151106" Sep 29 17:33:06 crc kubenswrapper[4667]: E0929 17:33:06.817084 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:33:10 crc kubenswrapper[4667]: I0929 17:33:10.314737 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:10 crc kubenswrapper[4667]: I0929 17:33:10.314972 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:10 crc kubenswrapper[4667]: I0929 17:33:10.344177 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:11 crc kubenswrapper[4667]: I0929 17:33:11.147047 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:11 crc kubenswrapper[4667]: I0929 17:33:11.174777 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rtfrc"] Sep 29 17:33:12 crc kubenswrapper[4667]: I0929 17:33:12.970432 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h5szb"] Sep 29 17:33:12 crc kubenswrapper[4667]: I0929 17:33:12.971982 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:12 crc kubenswrapper[4667]: I0929 17:33:12.978037 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h5szb"] Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.029931 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc67m\" (UniqueName: \"kubernetes.io/projected/07ade923-cf34-48b6-bced-91d203e9c363-kube-api-access-qc67m\") pod \"certified-operators-h5szb\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.030027 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-utilities\") pod \"certified-operators-h5szb\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.030045 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-catalog-content\") pod \"certified-operators-h5szb\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.127991 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rtfrc" podUID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerName="registry-server" containerID="cri-o://d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61" gracePeriod=2 Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.131253 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-utilities\") pod \"certified-operators-h5szb\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.131288 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-catalog-content\") pod \"certified-operators-h5szb\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.131443 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc67m\" (UniqueName: \"kubernetes.io/projected/07ade923-cf34-48b6-bced-91d203e9c363-kube-api-access-qc67m\") pod \"certified-operators-h5szb\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.132061 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-utilities\") pod \"certified-operators-h5szb\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.132130 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-catalog-content\") pod \"certified-operators-h5szb\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.147710 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc67m\" (UniqueName: \"kubernetes.io/projected/07ade923-cf34-48b6-bced-91d203e9c363-kube-api-access-qc67m\") pod \"certified-operators-h5szb\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.291817 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.659420 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h5szb"] Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.916833 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.942397 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-utilities\") pod \"9c7d713d-a85d-4a36-a338-d35fcb516b24\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.942428 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-catalog-content\") pod \"9c7d713d-a85d-4a36-a338-d35fcb516b24\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.942578 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg987\" (UniqueName: \"kubernetes.io/projected/9c7d713d-a85d-4a36-a338-d35fcb516b24-kube-api-access-hg987\") pod \"9c7d713d-a85d-4a36-a338-d35fcb516b24\" (UID: \"9c7d713d-a85d-4a36-a338-d35fcb516b24\") " Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.942922 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-utilities" (OuterVolumeSpecName: "utilities") pod "9c7d713d-a85d-4a36-a338-d35fcb516b24" (UID: "9c7d713d-a85d-4a36-a338-d35fcb516b24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.947193 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c7d713d-a85d-4a36-a338-d35fcb516b24-kube-api-access-hg987" (OuterVolumeSpecName: "kube-api-access-hg987") pod "9c7d713d-a85d-4a36-a338-d35fcb516b24" (UID: "9c7d713d-a85d-4a36-a338-d35fcb516b24"). InnerVolumeSpecName "kube-api-access-hg987". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:33:13 crc kubenswrapper[4667]: I0929 17:33:13.997111 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c7d713d-a85d-4a36-a338-d35fcb516b24" (UID: "9c7d713d-a85d-4a36-a338-d35fcb516b24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.043915 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg987\" (UniqueName: \"kubernetes.io/projected/9c7d713d-a85d-4a36-a338-d35fcb516b24-kube-api-access-hg987\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.043941 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.043951 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c7d713d-a85d-4a36-a338-d35fcb516b24-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.135009 4667 generic.go:334] "Generic (PLEG): container finished" podID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerID="d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61" exitCode=0 Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.135037 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rtfrc" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.135071 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtfrc" event={"ID":"9c7d713d-a85d-4a36-a338-d35fcb516b24","Type":"ContainerDied","Data":"d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61"} Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.135105 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rtfrc" event={"ID":"9c7d713d-a85d-4a36-a338-d35fcb516b24","Type":"ContainerDied","Data":"5dc4ad4be77d1151e759971efebcad8b9d2e2ac4e0d2165283da31a40182c649"} Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.135121 4667 scope.go:117] "RemoveContainer" containerID="d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.136498 4667 generic.go:334] "Generic (PLEG): container finished" podID="07ade923-cf34-48b6-bced-91d203e9c363" containerID="ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f" exitCode=0 Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.136535 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5szb" event={"ID":"07ade923-cf34-48b6-bced-91d203e9c363","Type":"ContainerDied","Data":"ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f"} Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.136559 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5szb" event={"ID":"07ade923-cf34-48b6-bced-91d203e9c363","Type":"ContainerStarted","Data":"53fa65522a2682986792252199698ab3f6d7dcf329ab0e0007ecc8e6468beccb"} Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.150560 4667 scope.go:117] "RemoveContainer" containerID="44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.165316 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rtfrc"] Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.169511 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rtfrc"] Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.181289 4667 scope.go:117] "RemoveContainer" containerID="8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.198775 4667 scope.go:117] "RemoveContainer" containerID="d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61" Sep 29 17:33:14 crc kubenswrapper[4667]: E0929 17:33:14.199252 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61\": container with ID starting with d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61 not found: ID does not exist" containerID="d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.199298 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61"} err="failed to get container status \"d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61\": rpc error: code = NotFound desc = could not find container \"d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61\": container with ID starting with d8d8077bed5f72cc94201bb3d47919df4a27010db205891efdbd83838b877e61 not found: ID does not exist" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.199323 4667 scope.go:117] "RemoveContainer" containerID="44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1" Sep 29 17:33:14 crc kubenswrapper[4667]: E0929 17:33:14.199596 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1\": container with ID starting with 44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1 not found: ID does not exist" containerID="44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.199623 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1"} err="failed to get container status \"44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1\": rpc error: code = NotFound desc = could not find container \"44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1\": container with ID starting with 44b66a6407b512986b6f417367a8b1462df0a1e6704a9445bdf5805824754ff1 not found: ID does not exist" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.199640 4667 scope.go:117] "RemoveContainer" containerID="8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022" Sep 29 17:33:14 crc kubenswrapper[4667]: E0929 17:33:14.199892 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022\": container with ID starting with 8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022 not found: ID does not exist" containerID="8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022" Sep 29 17:33:14 crc kubenswrapper[4667]: I0929 17:33:14.199916 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022"} err="failed to get container status \"8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022\": rpc error: code = NotFound desc = could not find container \"8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022\": container with ID starting with 8b4c0ceb4f295f233d3b99c49224280b6fd0cbeb860785426c91d8488c332022 not found: ID does not exist" Sep 29 17:33:15 crc kubenswrapper[4667]: I0929 17:33:15.144045 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5szb" event={"ID":"07ade923-cf34-48b6-bced-91d203e9c363","Type":"ContainerStarted","Data":"f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216"} Sep 29 17:33:15 crc kubenswrapper[4667]: I0929 17:33:15.822509 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c7d713d-a85d-4a36-a338-d35fcb516b24" path="/var/lib/kubelet/pods/9c7d713d-a85d-4a36-a338-d35fcb516b24/volumes" Sep 29 17:33:16 crc kubenswrapper[4667]: I0929 17:33:16.150386 4667 generic.go:334] "Generic (PLEG): container finished" podID="07ade923-cf34-48b6-bced-91d203e9c363" containerID="f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216" exitCode=0 Sep 29 17:33:16 crc kubenswrapper[4667]: I0929 17:33:16.150422 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5szb" event={"ID":"07ade923-cf34-48b6-bced-91d203e9c363","Type":"ContainerDied","Data":"f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216"} Sep 29 17:33:17 crc kubenswrapper[4667]: I0929 17:33:17.156978 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5szb" event={"ID":"07ade923-cf34-48b6-bced-91d203e9c363","Type":"ContainerStarted","Data":"aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a"} Sep 29 17:33:17 crc kubenswrapper[4667]: I0929 17:33:17.168763 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h5szb" podStartSLOduration=2.6023171339999998 podStartE2EDuration="5.168746361s" podCreationTimestamp="2025-09-29 17:33:12 +0000 UTC" firstStartedPulling="2025-09-29 17:33:14.137893569 +0000 UTC m=+1422.635740338" lastFinishedPulling="2025-09-29 17:33:16.704322806 +0000 UTC m=+1425.202169565" observedRunningTime="2025-09-29 17:33:17.167370328 +0000 UTC m=+1425.665217097" watchObservedRunningTime="2025-09-29 17:33:17.168746361 +0000 UTC m=+1425.666593130" Sep 29 17:33:17 crc kubenswrapper[4667]: E0929 17:33:17.816963 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.171953 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2zdv2"] Sep 29 17:33:21 crc kubenswrapper[4667]: E0929 17:33:21.172277 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerName="extract-utilities" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.172291 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerName="extract-utilities" Sep 29 17:33:21 crc kubenswrapper[4667]: E0929 17:33:21.172320 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerName="extract-content" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.172327 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerName="extract-content" Sep 29 17:33:21 crc kubenswrapper[4667]: E0929 17:33:21.172337 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerName="registry-server" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.172343 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerName="registry-server" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.172451 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c7d713d-a85d-4a36-a338-d35fcb516b24" containerName="registry-server" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.173459 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.177950 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zdv2"] Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.227738 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2gsx\" (UniqueName: \"kubernetes.io/projected/affc5191-8103-4c63-824d-b1e3c029b5ad-kube-api-access-n2gsx\") pod \"redhat-marketplace-2zdv2\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.227966 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-catalog-content\") pod \"redhat-marketplace-2zdv2\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.228004 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-utilities\") pod \"redhat-marketplace-2zdv2\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.329120 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-catalog-content\") pod \"redhat-marketplace-2zdv2\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.329164 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-utilities\") pod \"redhat-marketplace-2zdv2\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.329205 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2gsx\" (UniqueName: \"kubernetes.io/projected/affc5191-8103-4c63-824d-b1e3c029b5ad-kube-api-access-n2gsx\") pod \"redhat-marketplace-2zdv2\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.329771 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-catalog-content\") pod \"redhat-marketplace-2zdv2\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.329804 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-utilities\") pod \"redhat-marketplace-2zdv2\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.344053 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2gsx\" (UniqueName: \"kubernetes.io/projected/affc5191-8103-4c63-824d-b1e3c029b5ad-kube-api-access-n2gsx\") pod \"redhat-marketplace-2zdv2\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.375677 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rpp26"] Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.376984 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.382351 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpp26"] Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.430596 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpncw\" (UniqueName: \"kubernetes.io/projected/2aa01f4d-8a2a-4d6e-8681-d63497519357-kube-api-access-jpncw\") pod \"community-operators-rpp26\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.430655 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-catalog-content\") pod \"community-operators-rpp26\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.430742 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-utilities\") pod \"community-operators-rpp26\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.488801 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.531725 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpncw\" (UniqueName: \"kubernetes.io/projected/2aa01f4d-8a2a-4d6e-8681-d63497519357-kube-api-access-jpncw\") pod \"community-operators-rpp26\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.531778 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-catalog-content\") pod \"community-operators-rpp26\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.531814 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-utilities\") pod \"community-operators-rpp26\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.532533 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-catalog-content\") pod \"community-operators-rpp26\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.532541 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-utilities\") pod \"community-operators-rpp26\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.546552 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpncw\" (UniqueName: \"kubernetes.io/projected/2aa01f4d-8a2a-4d6e-8681-d63497519357-kube-api-access-jpncw\") pod \"community-operators-rpp26\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.697449 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:21 crc kubenswrapper[4667]: I0929 17:33:21.884176 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zdv2"] Sep 29 17:33:21 crc kubenswrapper[4667]: W0929 17:33:21.885586 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaffc5191_8103_4c63_824d_b1e3c029b5ad.slice/crio-36138d4b8164bc70e88fa9420e9d0475269e42cf50a70f9c6eb2bed5cb1f41ce WatchSource:0}: Error finding container 36138d4b8164bc70e88fa9420e9d0475269e42cf50a70f9c6eb2bed5cb1f41ce: Status 404 returned error can't find the container with id 36138d4b8164bc70e88fa9420e9d0475269e42cf50a70f9c6eb2bed5cb1f41ce Sep 29 17:33:22 crc kubenswrapper[4667]: I0929 17:33:22.055976 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpp26"] Sep 29 17:33:22 crc kubenswrapper[4667]: W0929 17:33:22.068083 4667 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2aa01f4d_8a2a_4d6e_8681_d63497519357.slice/crio-86bf00953bc064bf648c4be7e3c05998822f50fbddb77eacb567e92fbde4beaf WatchSource:0}: Error finding container 86bf00953bc064bf648c4be7e3c05998822f50fbddb77eacb567e92fbde4beaf: Status 404 returned error can't find the container with id 86bf00953bc064bf648c4be7e3c05998822f50fbddb77eacb567e92fbde4beaf Sep 29 17:33:22 crc kubenswrapper[4667]: I0929 17:33:22.185665 4667 generic.go:334] "Generic (PLEG): container finished" podID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerID="717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df" exitCode=0 Sep 29 17:33:22 crc kubenswrapper[4667]: I0929 17:33:22.185703 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zdv2" event={"ID":"affc5191-8103-4c63-824d-b1e3c029b5ad","Type":"ContainerDied","Data":"717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df"} Sep 29 17:33:22 crc kubenswrapper[4667]: I0929 17:33:22.185745 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zdv2" event={"ID":"affc5191-8103-4c63-824d-b1e3c029b5ad","Type":"ContainerStarted","Data":"36138d4b8164bc70e88fa9420e9d0475269e42cf50a70f9c6eb2bed5cb1f41ce"} Sep 29 17:33:22 crc kubenswrapper[4667]: I0929 17:33:22.187068 4667 generic.go:334] "Generic (PLEG): container finished" podID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerID="c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622" exitCode=0 Sep 29 17:33:22 crc kubenswrapper[4667]: I0929 17:33:22.187105 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp26" event={"ID":"2aa01f4d-8a2a-4d6e-8681-d63497519357","Type":"ContainerDied","Data":"c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622"} Sep 29 17:33:22 crc kubenswrapper[4667]: I0929 17:33:22.187128 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp26" event={"ID":"2aa01f4d-8a2a-4d6e-8681-d63497519357","Type":"ContainerStarted","Data":"86bf00953bc064bf648c4be7e3c05998822f50fbddb77eacb567e92fbde4beaf"} Sep 29 17:33:23 crc kubenswrapper[4667]: I0929 17:33:23.193702 4667 generic.go:334] "Generic (PLEG): container finished" podID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerID="621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce" exitCode=0 Sep 29 17:33:23 crc kubenswrapper[4667]: I0929 17:33:23.193872 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp26" event={"ID":"2aa01f4d-8a2a-4d6e-8681-d63497519357","Type":"ContainerDied","Data":"621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce"} Sep 29 17:33:23 crc kubenswrapper[4667]: I0929 17:33:23.196056 4667 generic.go:334] "Generic (PLEG): container finished" podID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerID="5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906" exitCode=0 Sep 29 17:33:23 crc kubenswrapper[4667]: I0929 17:33:23.196092 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zdv2" event={"ID":"affc5191-8103-4c63-824d-b1e3c029b5ad","Type":"ContainerDied","Data":"5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906"} Sep 29 17:33:23 crc kubenswrapper[4667]: I0929 17:33:23.292401 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:23 crc kubenswrapper[4667]: I0929 17:33:23.292438 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:23 crc kubenswrapper[4667]: I0929 17:33:23.321260 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:24 crc kubenswrapper[4667]: I0929 17:33:24.202939 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp26" event={"ID":"2aa01f4d-8a2a-4d6e-8681-d63497519357","Type":"ContainerStarted","Data":"725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81"} Sep 29 17:33:24 crc kubenswrapper[4667]: I0929 17:33:24.204791 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zdv2" event={"ID":"affc5191-8103-4c63-824d-b1e3c029b5ad","Type":"ContainerStarted","Data":"600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c"} Sep 29 17:33:24 crc kubenswrapper[4667]: I0929 17:33:24.234398 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rpp26" podStartSLOduration=1.677184287 podStartE2EDuration="3.234383517s" podCreationTimestamp="2025-09-29 17:33:21 +0000 UTC" firstStartedPulling="2025-09-29 17:33:22.188207958 +0000 UTC m=+1430.686054726" lastFinishedPulling="2025-09-29 17:33:23.745407187 +0000 UTC m=+1432.243253956" observedRunningTime="2025-09-29 17:33:24.221357125 +0000 UTC m=+1432.719203894" watchObservedRunningTime="2025-09-29 17:33:24.234383517 +0000 UTC m=+1432.732230286" Sep 29 17:33:24 crc kubenswrapper[4667]: I0929 17:33:24.237068 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:24 crc kubenswrapper[4667]: I0929 17:33:24.253450 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2zdv2" podStartSLOduration=1.6204016 podStartE2EDuration="3.253436345s" podCreationTimestamp="2025-09-29 17:33:21 +0000 UTC" firstStartedPulling="2025-09-29 17:33:22.186852543 +0000 UTC m=+1430.684699312" lastFinishedPulling="2025-09-29 17:33:23.819887289 +0000 UTC m=+1432.317734057" observedRunningTime="2025-09-29 17:33:24.234920228 +0000 UTC m=+1432.732766998" watchObservedRunningTime="2025-09-29 17:33:24.253436345 +0000 UTC m=+1432.751283115" Sep 29 17:33:25 crc kubenswrapper[4667]: I0929 17:33:25.766713 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h5szb"] Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.218240 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h5szb" podUID="07ade923-cf34-48b6-bced-91d203e9c363" containerName="registry-server" containerID="cri-o://aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a" gracePeriod=2 Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.529605 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.697170 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-utilities\") pod \"07ade923-cf34-48b6-bced-91d203e9c363\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.697206 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-catalog-content\") pod \"07ade923-cf34-48b6-bced-91d203e9c363\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.697341 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc67m\" (UniqueName: \"kubernetes.io/projected/07ade923-cf34-48b6-bced-91d203e9c363-kube-api-access-qc67m\") pod \"07ade923-cf34-48b6-bced-91d203e9c363\" (UID: \"07ade923-cf34-48b6-bced-91d203e9c363\") " Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.697676 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-utilities" (OuterVolumeSpecName: "utilities") pod "07ade923-cf34-48b6-bced-91d203e9c363" (UID: "07ade923-cf34-48b6-bced-91d203e9c363"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.701527 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07ade923-cf34-48b6-bced-91d203e9c363-kube-api-access-qc67m" (OuterVolumeSpecName: "kube-api-access-qc67m") pod "07ade923-cf34-48b6-bced-91d203e9c363" (UID: "07ade923-cf34-48b6-bced-91d203e9c363"). InnerVolumeSpecName "kube-api-access-qc67m". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.725652 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07ade923-cf34-48b6-bced-91d203e9c363" (UID: "07ade923-cf34-48b6-bced-91d203e9c363"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.798772 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc67m\" (UniqueName: \"kubernetes.io/projected/07ade923-cf34-48b6-bced-91d203e9c363-kube-api-access-qc67m\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.798803 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.799174 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07ade923-cf34-48b6-bced-91d203e9c363-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.897022 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:33:26 crc kubenswrapper[4667]: I0929 17:33:26.897072 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.224681 4667 generic.go:334] "Generic (PLEG): container finished" podID="07ade923-cf34-48b6-bced-91d203e9c363" containerID="aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a" exitCode=0 Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.224718 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5szb" event={"ID":"07ade923-cf34-48b6-bced-91d203e9c363","Type":"ContainerDied","Data":"aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a"} Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.224744 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h5szb" event={"ID":"07ade923-cf34-48b6-bced-91d203e9c363","Type":"ContainerDied","Data":"53fa65522a2682986792252199698ab3f6d7dcf329ab0e0007ecc8e6468beccb"} Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.224759 4667 scope.go:117] "RemoveContainer" containerID="aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.224877 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h5szb" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.238437 4667 scope.go:117] "RemoveContainer" containerID="f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.246094 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h5szb"] Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.250471 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h5szb"] Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.251894 4667 scope.go:117] "RemoveContainer" containerID="ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.271302 4667 scope.go:117] "RemoveContainer" containerID="aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a" Sep 29 17:33:27 crc kubenswrapper[4667]: E0929 17:33:27.271627 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a\": container with ID starting with aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a not found: ID does not exist" containerID="aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.271677 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a"} err="failed to get container status \"aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a\": rpc error: code = NotFound desc = could not find container \"aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a\": container with ID starting with aecefecc30db89c0ee515fa92f74fa7ca6e0588db2947c77f0977b87826e5f0a not found: ID does not exist" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.271696 4667 scope.go:117] "RemoveContainer" containerID="f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216" Sep 29 17:33:27 crc kubenswrapper[4667]: E0929 17:33:27.271997 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216\": container with ID starting with f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216 not found: ID does not exist" containerID="f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.272032 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216"} err="failed to get container status \"f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216\": rpc error: code = NotFound desc = could not find container \"f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216\": container with ID starting with f5dc15af9a45a4ccb1afcf8475b380a6de64e3552b50d524bb9740c38c0e4216 not found: ID does not exist" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.272054 4667 scope.go:117] "RemoveContainer" containerID="ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f" Sep 29 17:33:27 crc kubenswrapper[4667]: E0929 17:33:27.272260 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f\": container with ID starting with ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f not found: ID does not exist" containerID="ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.272280 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f"} err="failed to get container status \"ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f\": rpc error: code = NotFound desc = could not find container \"ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f\": container with ID starting with ae203e18428270aabaa5ebe195427a053a4f34f7e1d235e2e29a0c1d15ecde2f not found: ID does not exist" Sep 29 17:33:27 crc kubenswrapper[4667]: I0929 17:33:27.822389 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07ade923-cf34-48b6-bced-91d203e9c363" path="/var/lib/kubelet/pods/07ade923-cf34-48b6-bced-91d203e9c363/volumes" Sep 29 17:33:31 crc kubenswrapper[4667]: I0929 17:33:31.489031 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:31 crc kubenswrapper[4667]: I0929 17:33:31.489330 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:31 crc kubenswrapper[4667]: I0929 17:33:31.518072 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:31 crc kubenswrapper[4667]: I0929 17:33:31.697984 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:31 crc kubenswrapper[4667]: I0929 17:33:31.698020 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:31 crc kubenswrapper[4667]: I0929 17:33:31.727048 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:32 crc kubenswrapper[4667]: I0929 17:33:32.280614 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:32 crc kubenswrapper[4667]: I0929 17:33:32.280814 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:33 crc kubenswrapper[4667]: I0929 17:33:33.165960 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rpp26"] Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.267898 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rpp26" podUID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerName="registry-server" containerID="cri-o://725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81" gracePeriod=2 Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.568062 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zdv2"] Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.568392 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2zdv2" podUID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerName="registry-server" containerID="cri-o://600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c" gracePeriod=2 Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.669924 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.797875 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpncw\" (UniqueName: \"kubernetes.io/projected/2aa01f4d-8a2a-4d6e-8681-d63497519357-kube-api-access-jpncw\") pod \"2aa01f4d-8a2a-4d6e-8681-d63497519357\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.797947 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-catalog-content\") pod \"2aa01f4d-8a2a-4d6e-8681-d63497519357\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.798003 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-utilities\") pod \"2aa01f4d-8a2a-4d6e-8681-d63497519357\" (UID: \"2aa01f4d-8a2a-4d6e-8681-d63497519357\") " Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.798660 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-utilities" (OuterVolumeSpecName: "utilities") pod "2aa01f4d-8a2a-4d6e-8681-d63497519357" (UID: "2aa01f4d-8a2a-4d6e-8681-d63497519357"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.802369 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa01f4d-8a2a-4d6e-8681-d63497519357-kube-api-access-jpncw" (OuterVolumeSpecName: "kube-api-access-jpncw") pod "2aa01f4d-8a2a-4d6e-8681-d63497519357" (UID: "2aa01f4d-8a2a-4d6e-8681-d63497519357"). InnerVolumeSpecName "kube-api-access-jpncw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.832400 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2aa01f4d-8a2a-4d6e-8681-d63497519357" (UID: "2aa01f4d-8a2a-4d6e-8681-d63497519357"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.860674 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.899749 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpncw\" (UniqueName: \"kubernetes.io/projected/2aa01f4d-8a2a-4d6e-8681-d63497519357-kube-api-access-jpncw\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.899775 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:34 crc kubenswrapper[4667]: I0929 17:33:34.899784 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2aa01f4d-8a2a-4d6e-8681-d63497519357-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.001142 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-catalog-content\") pod \"affc5191-8103-4c63-824d-b1e3c029b5ad\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.001191 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-utilities\") pod \"affc5191-8103-4c63-824d-b1e3c029b5ad\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.001297 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2gsx\" (UniqueName: \"kubernetes.io/projected/affc5191-8103-4c63-824d-b1e3c029b5ad-kube-api-access-n2gsx\") pod \"affc5191-8103-4c63-824d-b1e3c029b5ad\" (UID: \"affc5191-8103-4c63-824d-b1e3c029b5ad\") " Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.001836 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-utilities" (OuterVolumeSpecName: "utilities") pod "affc5191-8103-4c63-824d-b1e3c029b5ad" (UID: "affc5191-8103-4c63-824d-b1e3c029b5ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.004055 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/affc5191-8103-4c63-824d-b1e3c029b5ad-kube-api-access-n2gsx" (OuterVolumeSpecName: "kube-api-access-n2gsx") pod "affc5191-8103-4c63-824d-b1e3c029b5ad" (UID: "affc5191-8103-4c63-824d-b1e3c029b5ad"). InnerVolumeSpecName "kube-api-access-n2gsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.010260 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "affc5191-8103-4c63-824d-b1e3c029b5ad" (UID: "affc5191-8103-4c63-824d-b1e3c029b5ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.103080 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.103107 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/affc5191-8103-4c63-824d-b1e3c029b5ad-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.103117 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2gsx\" (UniqueName: \"kubernetes.io/projected/affc5191-8103-4c63-824d-b1e3c029b5ad-kube-api-access-n2gsx\") on node \"crc\" DevicePath \"\"" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.274753 4667 generic.go:334] "Generic (PLEG): container finished" podID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerID="600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c" exitCode=0 Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.274824 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2zdv2" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.274863 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zdv2" event={"ID":"affc5191-8103-4c63-824d-b1e3c029b5ad","Type":"ContainerDied","Data":"600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c"} Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.274911 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2zdv2" event={"ID":"affc5191-8103-4c63-824d-b1e3c029b5ad","Type":"ContainerDied","Data":"36138d4b8164bc70e88fa9420e9d0475269e42cf50a70f9c6eb2bed5cb1f41ce"} Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.274925 4667 scope.go:117] "RemoveContainer" containerID="600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.276991 4667 generic.go:334] "Generic (PLEG): container finished" podID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerID="725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81" exitCode=0 Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.277030 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp26" event={"ID":"2aa01f4d-8a2a-4d6e-8681-d63497519357","Type":"ContainerDied","Data":"725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81"} Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.277054 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpp26" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.277079 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpp26" event={"ID":"2aa01f4d-8a2a-4d6e-8681-d63497519357","Type":"ContainerDied","Data":"86bf00953bc064bf648c4be7e3c05998822f50fbddb77eacb567e92fbde4beaf"} Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.288630 4667 scope.go:117] "RemoveContainer" containerID="5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.298872 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zdv2"] Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.304429 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2zdv2"] Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.305341 4667 scope.go:117] "RemoveContainer" containerID="717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.308294 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rpp26"] Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.312566 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rpp26"] Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.329778 4667 scope.go:117] "RemoveContainer" containerID="600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c" Sep 29 17:33:35 crc kubenswrapper[4667]: E0929 17:33:35.330038 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c\": container with ID starting with 600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c not found: ID does not exist" containerID="600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.330069 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c"} err="failed to get container status \"600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c\": rpc error: code = NotFound desc = could not find container \"600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c\": container with ID starting with 600c2835d409142e9fb9f0fd61ab098197e38bee799782e394800da8a41d7a7c not found: ID does not exist" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.330091 4667 scope.go:117] "RemoveContainer" containerID="5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906" Sep 29 17:33:35 crc kubenswrapper[4667]: E0929 17:33:35.330339 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906\": container with ID starting with 5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906 not found: ID does not exist" containerID="5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.330365 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906"} err="failed to get container status \"5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906\": rpc error: code = NotFound desc = could not find container \"5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906\": container with ID starting with 5f5ca448e6733b46d031f72f5b8f3947d11f57288a13fde5a9ab21a80391f906 not found: ID does not exist" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.330383 4667 scope.go:117] "RemoveContainer" containerID="717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df" Sep 29 17:33:35 crc kubenswrapper[4667]: E0929 17:33:35.330649 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df\": container with ID starting with 717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df not found: ID does not exist" containerID="717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.330670 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df"} err="failed to get container status \"717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df\": rpc error: code = NotFound desc = could not find container \"717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df\": container with ID starting with 717c078653b70473b3ad3bf0b34ff26397e098055594a13c6224ee7d56cd33df not found: ID does not exist" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.330682 4667 scope.go:117] "RemoveContainer" containerID="725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.342122 4667 scope.go:117] "RemoveContainer" containerID="621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.355331 4667 scope.go:117] "RemoveContainer" containerID="c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.365813 4667 scope.go:117] "RemoveContainer" containerID="725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81" Sep 29 17:33:35 crc kubenswrapper[4667]: E0929 17:33:35.366203 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81\": container with ID starting with 725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81 not found: ID does not exist" containerID="725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.366231 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81"} err="failed to get container status \"725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81\": rpc error: code = NotFound desc = could not find container \"725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81\": container with ID starting with 725d06de3b2e55f24f00a34e2d26445fea6ecc7b772d2c9f93f1effc709afd81 not found: ID does not exist" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.366249 4667 scope.go:117] "RemoveContainer" containerID="621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce" Sep 29 17:33:35 crc kubenswrapper[4667]: E0929 17:33:35.366504 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce\": container with ID starting with 621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce not found: ID does not exist" containerID="621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.366547 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce"} err="failed to get container status \"621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce\": rpc error: code = NotFound desc = could not find container \"621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce\": container with ID starting with 621427b978510b5cbad5f266d5ad9a14b547a2e26dec57bc797b5e74521819ce not found: ID does not exist" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.366574 4667 scope.go:117] "RemoveContainer" containerID="c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622" Sep 29 17:33:35 crc kubenswrapper[4667]: E0929 17:33:35.366838 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622\": container with ID starting with c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622 not found: ID does not exist" containerID="c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.366877 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622"} err="failed to get container status \"c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622\": rpc error: code = NotFound desc = could not find container \"c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622\": container with ID starting with c143bf632b31737a92152144697fe1807e6f26022920f92937ebf0a20d2e0622 not found: ID does not exist" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.822862 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aa01f4d-8a2a-4d6e-8681-d63497519357" path="/var/lib/kubelet/pods/2aa01f4d-8a2a-4d6e-8681-d63497519357/volumes" Sep 29 17:33:35 crc kubenswrapper[4667]: I0929 17:33:35.823574 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="affc5191-8103-4c63-824d-b1e3c029b5ad" path="/var/lib/kubelet/pods/affc5191-8103-4c63-824d-b1e3c029b5ad/volumes" Sep 29 17:33:56 crc kubenswrapper[4667]: I0929 17:33:56.897405 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:33:56 crc kubenswrapper[4667]: I0929 17:33:56.897991 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:33:56 crc kubenswrapper[4667]: I0929 17:33:56.898051 4667 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:33:56 crc kubenswrapper[4667]: I0929 17:33:56.898961 4667 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9c9d598e95ffa725e3f0ee5b1aa348c0acb3951aa004315e545eed19b8702772"} pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 17:33:56 crc kubenswrapper[4667]: I0929 17:33:56.899017 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" containerID="cri-o://9c9d598e95ffa725e3f0ee5b1aa348c0acb3951aa004315e545eed19b8702772" gracePeriod=600 Sep 29 17:33:57 crc kubenswrapper[4667]: I0929 17:33:57.409800 4667 generic.go:334] "Generic (PLEG): container finished" podID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerID="9c9d598e95ffa725e3f0ee5b1aa348c0acb3951aa004315e545eed19b8702772" exitCode=0 Sep 29 17:33:57 crc kubenswrapper[4667]: I0929 17:33:57.409830 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerDied","Data":"9c9d598e95ffa725e3f0ee5b1aa348c0acb3951aa004315e545eed19b8702772"} Sep 29 17:33:57 crc kubenswrapper[4667]: I0929 17:33:57.410011 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861"} Sep 29 17:33:57 crc kubenswrapper[4667]: I0929 17:33:57.410030 4667 scope.go:117] "RemoveContainer" containerID="94b33344ce4f405cbd10df51f10373f90996afec3b5c19edb1480b7527aa57d4" Sep 29 17:35:28 crc kubenswrapper[4667]: E0929 17:35:28.822225 4667 log.go:32] "PullImage from image service failed" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:35:28 crc kubenswrapper[4667]: E0929 17:35:28.822752 4667 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:35:28 crc kubenswrapper[4667]: E0929 17:35:28.822897 4667 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmkl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-operator-index-6b7qn_openstack-operators(7aa88991-5ef2-4166-8d37-8ce5fb207d26): ErrImagePull: rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" logger="UnhandledError" Sep 29 17:35:28 crc kubenswrapper[4667]: E0929 17:35:28.824069 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \\\"http://38.129.56.34:5001/v2/\\\": dial tcp 38.129.56.34:5001: i/o timeout\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:35:40 crc kubenswrapper[4667]: E0929 17:35:40.816982 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:35:54 crc kubenswrapper[4667]: E0929 17:35:54.816693 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:36:05 crc kubenswrapper[4667]: E0929 17:36:05.817580 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:36:20 crc kubenswrapper[4667]: E0929 17:36:20.817291 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:36:26 crc kubenswrapper[4667]: I0929 17:36:26.896923 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:36:26 crc kubenswrapper[4667]: I0929 17:36:26.897090 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:36:34 crc kubenswrapper[4667]: E0929 17:36:34.817057 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:36:56 crc kubenswrapper[4667]: I0929 17:36:56.897885 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:36:56 crc kubenswrapper[4667]: I0929 17:36:56.898819 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:37:26 crc kubenswrapper[4667]: I0929 17:37:26.897517 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:37:26 crc kubenswrapper[4667]: I0929 17:37:26.898155 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:37:26 crc kubenswrapper[4667]: I0929 17:37:26.898203 4667 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" Sep 29 17:37:26 crc kubenswrapper[4667]: I0929 17:37:26.898741 4667 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861"} pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Sep 29 17:37:26 crc kubenswrapper[4667]: I0929 17:37:26.898789 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" containerID="cri-o://244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" gracePeriod=600 Sep 29 17:37:27 crc kubenswrapper[4667]: E0929 17:37:27.017050 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:37:27 crc kubenswrapper[4667]: I0929 17:37:27.635941 4667 generic.go:334] "Generic (PLEG): container finished" podID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" exitCode=0 Sep 29 17:37:27 crc kubenswrapper[4667]: I0929 17:37:27.635989 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerDied","Data":"244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861"} Sep 29 17:37:27 crc kubenswrapper[4667]: I0929 17:37:27.636028 4667 scope.go:117] "RemoveContainer" containerID="9c9d598e95ffa725e3f0ee5b1aa348c0acb3951aa004315e545eed19b8702772" Sep 29 17:37:27 crc kubenswrapper[4667]: I0929 17:37:27.636613 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:37:27 crc kubenswrapper[4667]: E0929 17:37:27.636886 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:37:40 crc kubenswrapper[4667]: I0929 17:37:40.815563 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:37:40 crc kubenswrapper[4667]: E0929 17:37:40.816405 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:37:55 crc kubenswrapper[4667]: I0929 17:37:55.816408 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:37:55 crc kubenswrapper[4667]: E0929 17:37:55.817098 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:38:07 crc kubenswrapper[4667]: I0929 17:38:07.817078 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:38:07 crc kubenswrapper[4667]: E0929 17:38:07.817883 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:38:20 crc kubenswrapper[4667]: I0929 17:38:20.816226 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:38:20 crc kubenswrapper[4667]: E0929 17:38:20.817038 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:38:33 crc kubenswrapper[4667]: I0929 17:38:33.816325 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:38:33 crc kubenswrapper[4667]: E0929 17:38:33.817061 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:38:45 crc kubenswrapper[4667]: I0929 17:38:45.815428 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:38:45 crc kubenswrapper[4667]: E0929 17:38:45.816637 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:38:49 crc kubenswrapper[4667]: E0929 17:38:49.821024 4667 log.go:32] "PullImage from image service failed" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:38:49 crc kubenswrapper[4667]: E0929 17:38:49.821655 4667 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:38:49 crc kubenswrapper[4667]: E0929 17:38:49.821820 4667 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmkl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-operator-index-6b7qn_openstack-operators(7aa88991-5ef2-4166-8d37-8ce5fb207d26): ErrImagePull: rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" logger="UnhandledError" Sep 29 17:38:49 crc kubenswrapper[4667]: E0929 17:38:49.823467 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \\\"http://38.129.56.34:5001/v2/\\\": dial tcp 38.129.56.34:5001: i/o timeout\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:39:00 crc kubenswrapper[4667]: I0929 17:39:00.816070 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:39:00 crc kubenswrapper[4667]: E0929 17:39:00.816918 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:39:04 crc kubenswrapper[4667]: E0929 17:39:04.818332 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:39:15 crc kubenswrapper[4667]: I0929 17:39:15.816899 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:39:15 crc kubenswrapper[4667]: E0929 17:39:15.817950 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:39:16 crc kubenswrapper[4667]: E0929 17:39:16.819057 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:39:27 crc kubenswrapper[4667]: I0929 17:39:27.815896 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:39:27 crc kubenswrapper[4667]: E0929 17:39:27.816734 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:39:29 crc kubenswrapper[4667]: E0929 17:39:29.816997 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:39:41 crc kubenswrapper[4667]: I0929 17:39:41.820394 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:39:41 crc kubenswrapper[4667]: E0929 17:39:41.821194 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:39:42 crc kubenswrapper[4667]: E0929 17:39:42.818173 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:39:55 crc kubenswrapper[4667]: I0929 17:39:55.815467 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:39:55 crc kubenswrapper[4667]: E0929 17:39:55.816139 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:39:56 crc kubenswrapper[4667]: E0929 17:39:56.817462 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:40:10 crc kubenswrapper[4667]: I0929 17:40:10.815571 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:40:10 crc kubenswrapper[4667]: E0929 17:40:10.816490 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:40:11 crc kubenswrapper[4667]: E0929 17:40:11.819544 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:40:22 crc kubenswrapper[4667]: E0929 17:40:22.816568 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:40:25 crc kubenswrapper[4667]: I0929 17:40:25.815786 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:40:25 crc kubenswrapper[4667]: E0929 17:40:25.816350 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:40:33 crc kubenswrapper[4667]: E0929 17:40:33.817537 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:40:40 crc kubenswrapper[4667]: I0929 17:40:40.815983 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:40:40 crc kubenswrapper[4667]: E0929 17:40:40.816668 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:40:48 crc kubenswrapper[4667]: E0929 17:40:48.817363 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:40:55 crc kubenswrapper[4667]: I0929 17:40:55.815508 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:40:55 crc kubenswrapper[4667]: E0929 17:40:55.816196 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:41:02 crc kubenswrapper[4667]: E0929 17:41:02.816800 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:41:09 crc kubenswrapper[4667]: I0929 17:41:09.815473 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:41:09 crc kubenswrapper[4667]: E0929 17:41:09.816167 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:41:15 crc kubenswrapper[4667]: E0929 17:41:15.817061 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022173 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bns76/must-gather-m6926"] Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.022820 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerName="extract-content" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022833 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerName="extract-content" Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.022858 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerName="extract-utilities" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022865 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerName="extract-utilities" Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.022872 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerName="registry-server" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022877 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerName="registry-server" Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.022886 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerName="registry-server" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022891 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerName="registry-server" Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.022901 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ade923-cf34-48b6-bced-91d203e9c363" containerName="registry-server" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022907 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ade923-cf34-48b6-bced-91d203e9c363" containerName="registry-server" Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.022915 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerName="extract-utilities" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022921 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerName="extract-utilities" Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.022929 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ade923-cf34-48b6-bced-91d203e9c363" containerName="extract-content" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022934 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ade923-cf34-48b6-bced-91d203e9c363" containerName="extract-content" Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.022947 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ade923-cf34-48b6-bced-91d203e9c363" containerName="extract-utilities" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022952 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ade923-cf34-48b6-bced-91d203e9c363" containerName="extract-utilities" Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.022968 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerName="extract-content" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.022972 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerName="extract-content" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.023092 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="affc5191-8103-4c63-824d-b1e3c029b5ad" containerName="registry-server" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.023106 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="07ade923-cf34-48b6-bced-91d203e9c363" containerName="registry-server" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.023114 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa01f4d-8a2a-4d6e-8681-d63497519357" containerName="registry-server" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.023768 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.025949 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bns76"/"openshift-service-ca.crt" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.026127 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bns76"/"default-dockercfg-82xbf" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.026273 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bns76"/"kube-root-ca.crt" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.034513 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bns76/must-gather-m6926"] Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.099794 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/549699f7-0860-44fd-9bce-a37ce789396c-must-gather-output\") pod \"must-gather-m6926\" (UID: \"549699f7-0860-44fd-9bce-a37ce789396c\") " pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.099870 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsm4f\" (UniqueName: \"kubernetes.io/projected/549699f7-0860-44fd-9bce-a37ce789396c-kube-api-access-fsm4f\") pod \"must-gather-m6926\" (UID: \"549699f7-0860-44fd-9bce-a37ce789396c\") " pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.200856 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/549699f7-0860-44fd-9bce-a37ce789396c-must-gather-output\") pod \"must-gather-m6926\" (UID: \"549699f7-0860-44fd-9bce-a37ce789396c\") " pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.200918 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsm4f\" (UniqueName: \"kubernetes.io/projected/549699f7-0860-44fd-9bce-a37ce789396c-kube-api-access-fsm4f\") pod \"must-gather-m6926\" (UID: \"549699f7-0860-44fd-9bce-a37ce789396c\") " pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.201208 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/549699f7-0860-44fd-9bce-a37ce789396c-must-gather-output\") pod \"must-gather-m6926\" (UID: \"549699f7-0860-44fd-9bce-a37ce789396c\") " pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.216555 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsm4f\" (UniqueName: \"kubernetes.io/projected/549699f7-0860-44fd-9bce-a37ce789396c-kube-api-access-fsm4f\") pod \"must-gather-m6926\" (UID: \"549699f7-0860-44fd-9bce-a37ce789396c\") " pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.336629 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.724104 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bns76/must-gather-m6926"] Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.727871 4667 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Sep 29 17:41:23 crc kubenswrapper[4667]: I0929 17:41:23.815793 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:41:23 crc kubenswrapper[4667]: E0929 17:41:23.816066 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:41:24 crc kubenswrapper[4667]: I0929 17:41:24.089294 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bns76/must-gather-m6926" event={"ID":"549699f7-0860-44fd-9bce-a37ce789396c","Type":"ContainerStarted","Data":"4f8af251191878f4526355c1b3e1eee272abf02c4d215f27e44bc263e0bc6238"} Sep 29 17:41:26 crc kubenswrapper[4667]: E0929 17:41:26.817662 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:41:30 crc kubenswrapper[4667]: I0929 17:41:30.133770 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bns76/must-gather-m6926" event={"ID":"549699f7-0860-44fd-9bce-a37ce789396c","Type":"ContainerStarted","Data":"818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6"} Sep 29 17:41:30 crc kubenswrapper[4667]: I0929 17:41:30.134543 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bns76/must-gather-m6926" event={"ID":"549699f7-0860-44fd-9bce-a37ce789396c","Type":"ContainerStarted","Data":"e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54"} Sep 29 17:41:30 crc kubenswrapper[4667]: I0929 17:41:30.164974 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bns76/must-gather-m6926" podStartSLOduration=1.752890595 podStartE2EDuration="7.164954484s" podCreationTimestamp="2025-09-29 17:41:23 +0000 UTC" firstStartedPulling="2025-09-29 17:41:23.727637778 +0000 UTC m=+1912.225484548" lastFinishedPulling="2025-09-29 17:41:29.139701668 +0000 UTC m=+1917.637548437" observedRunningTime="2025-09-29 17:41:30.161063539 +0000 UTC m=+1918.658910308" watchObservedRunningTime="2025-09-29 17:41:30.164954484 +0000 UTC m=+1918.662801253" Sep 29 17:41:34 crc kubenswrapper[4667]: I0929 17:41:34.815239 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:41:34 crc kubenswrapper[4667]: E0929 17:41:34.816086 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:41:49 crc kubenswrapper[4667]: I0929 17:41:49.816293 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:41:49 crc kubenswrapper[4667]: E0929 17:41:49.817344 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:42:04 crc kubenswrapper[4667]: I0929 17:42:04.353440 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-7x9v7_3e396bd3-e718-4f53-a69b-522a601e0d4d/control-plane-machine-set-operator/0.log" Sep 29 17:42:04 crc kubenswrapper[4667]: I0929 17:42:04.435201 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j5pwf_5fb62347-e9c8-4815-a631-8fe0b5c78bd3/kube-rbac-proxy/0.log" Sep 29 17:42:04 crc kubenswrapper[4667]: I0929 17:42:04.467160 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-j5pwf_5fb62347-e9c8-4815-a631-8fe0b5c78bd3/machine-api-operator/0.log" Sep 29 17:42:04 crc kubenswrapper[4667]: I0929 17:42:04.816034 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:42:04 crc kubenswrapper[4667]: E0929 17:42:04.816300 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:42:14 crc kubenswrapper[4667]: I0929 17:42:14.754496 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-tx8qq_ce99b74e-e34b-4fbb-abfd-439f1c7f3585/cert-manager-controller/0.log" Sep 29 17:42:14 crc kubenswrapper[4667]: I0929 17:42:14.870816 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-qh727_77ca411f-8a0c-40b8-8457-57ed07672d2c/cert-manager-cainjector/0.log" Sep 29 17:42:14 crc kubenswrapper[4667]: I0929 17:42:14.941371 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-bzmsv_1eb84072-ac55-4869-a958-6b2419c839ae/cert-manager-webhook/0.log" Sep 29 17:42:16 crc kubenswrapper[4667]: I0929 17:42:16.815643 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:42:16 crc kubenswrapper[4667]: E0929 17:42:16.816253 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-l8rmj_openshift-machine-config-operator(28fa0016-3e75-4704-8b60-30ee9e576d59)\"" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" Sep 29 17:42:24 crc kubenswrapper[4667]: I0929 17:42:24.628994 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-864bb6dfb5-wzvhh_be8fb545-6b61-4464-b940-0fab07461beb/nmstate-console-plugin/0.log" Sep 29 17:42:24 crc kubenswrapper[4667]: I0929 17:42:24.770226 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-hcvkt_58bb8f72-34d0-44a3-a094-dc59c23fe49b/nmstate-handler/0.log" Sep 29 17:42:24 crc kubenswrapper[4667]: I0929 17:42:24.802761 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-nvhmg_b76f4f28-f053-48ff-bb61-524505148d8d/nmstate-metrics/0.log" Sep 29 17:42:24 crc kubenswrapper[4667]: I0929 17:42:24.802995 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-58fcddf996-nvhmg_b76f4f28-f053-48ff-bb61-524505148d8d/kube-rbac-proxy/0.log" Sep 29 17:42:24 crc kubenswrapper[4667]: I0929 17:42:24.930376 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5d6f6cfd66-k77n4_6d4c734e-b4e1-4daf-ba5f-24d5a4e2c163/nmstate-operator/0.log" Sep 29 17:42:24 crc kubenswrapper[4667]: I0929 17:42:24.961437 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6d689559c5-h5d7w_18dfc1a3-add5-49cb-860c-5361e887a24d/nmstate-webhook/0.log" Sep 29 17:42:29 crc kubenswrapper[4667]: I0929 17:42:29.819577 4667 scope.go:117] "RemoveContainer" containerID="244f80dfe4377ab388922cfe3e1c4f221f741d50b5221d2d4c431d86bb97e861" Sep 29 17:42:30 crc kubenswrapper[4667]: I0929 17:42:30.566698 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" event={"ID":"28fa0016-3e75-4704-8b60-30ee9e576d59","Type":"ContainerStarted","Data":"5a5e4d3a9908f3ecec9d5eef823be4f7554b724df09217a20438a11590c18ddc"} Sep 29 17:42:34 crc kubenswrapper[4667]: I0929 17:42:34.330800 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-679b68c759-tgpzd_fbde58f9-2869-40cd-bef8-a28336337936/kube-rbac-proxy/0.log" Sep 29 17:42:34 crc kubenswrapper[4667]: I0929 17:42:34.375874 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-679b68c759-tgpzd_fbde58f9-2869-40cd-bef8-a28336337936/manager/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.383948 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-fcc886d58-crc96_3519c647-b665-4a91-bcf2-3abbd6f982c9/cluster-logging-operator/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.533277 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-wvmfl_4bb000c8-f7ac-4c73-a717-0bd6bb93ca3a/collector/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.544294 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_647a9667-397f-4cb5-91bf-da76ac52ccad/loki-compactor/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.667627 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-55cc667c54-fm22d_f50c1454-bcbb-4a46-ab2c-cb717a1083d3/gateway/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.682420 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-67c9b4c785-twdch_d60533af-cf3f-4854-acd7-1aa7a11a47e9/loki-distributor/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.773598 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-55cc667c54-fm22d_f50c1454-bcbb-4a46-ab2c-cb717a1083d3/opa/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.821382 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-55cc667c54-mv8xl_defff6f8-6bee-4b88-b946-32cca2a46fc0/gateway/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.845662 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-55cc667c54-mv8xl_defff6f8-6bee-4b88-b946-32cca2a46fc0/opa/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.966577 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_980ab4d7-2edd-46f3-9f6c-b6138f1c3350/loki-index-gateway/0.log" Sep 29 17:42:44 crc kubenswrapper[4667]: I0929 17:42:44.993271 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_658afe18-7fa3-4e33-bf5c-cfab27925cb4/loki-ingester/0.log" Sep 29 17:42:45 crc kubenswrapper[4667]: I0929 17:42:45.109022 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-7454676c57-kv7bp_036ee610-9a8c-47ee-be84-32a681f82a61/loki-querier/0.log" Sep 29 17:42:45 crc kubenswrapper[4667]: I0929 17:42:45.174633 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-6b467cdd84-4479c_5d817e8c-29d5-48c6-8596-baddee83a3f3/loki-query-frontend/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.342831 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-gbfh9_c1afbef4-4d2f-4536-a4bc-5516962e56be/kube-rbac-proxy/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.482918 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5d688f5ffc-gbfh9_c1afbef4-4d2f-4536-a4bc-5516962e56be/controller/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.548933 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-frr-files/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.631136 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-frr-files/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.641571 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-metrics/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.656223 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-reloader/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.707969 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-reloader/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.829814 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-metrics/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.830000 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-frr-files/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.830060 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-reloader/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.870324 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-metrics/0.log" Sep 29 17:42:55 crc kubenswrapper[4667]: I0929 17:42:55.991552 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-frr-files/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.012932 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-reloader/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.019327 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/cp-metrics/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.029762 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/controller/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.154818 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/frr-metrics/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.167967 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/kube-rbac-proxy/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.214657 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/kube-rbac-proxy-frr/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.231331 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/frr/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.325230 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-tbwpr_d41412a0-9af7-438a-b70d-a391bd35670d/reloader/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.406730 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-5478bdb765-xmrg9_3e4ad89e-146c-4c00-b666-1ebf9487ef3a/frr-k8s-webhook-server/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.482133 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-67478f5d6-55rdc_f632e0ed-f363-40ef-a50e-ba37e1fad5c6/manager/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.558138 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-84bd8cf5c8-df27m_22444915-afb5-4f7b-92e4-b33472abb68e/webhook-server/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.663260 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wwv5m_8d9654e1-f7cf-4e80-95a8-9a133ca2e231/kube-rbac-proxy/0.log" Sep 29 17:42:56 crc kubenswrapper[4667]: I0929 17:42:56.791250 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wwv5m_8d9654e1-f7cf-4e80-95a8-9a133ca2e231/speaker/0.log" Sep 29 17:43:05 crc kubenswrapper[4667]: I0929 17:43:05.614046 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll_19f4040e-c856-4d41-a9a0-977dbce39c98/util/0.log" Sep 29 17:43:05 crc kubenswrapper[4667]: I0929 17:43:05.794609 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll_19f4040e-c856-4d41-a9a0-977dbce39c98/pull/0.log" Sep 29 17:43:05 crc kubenswrapper[4667]: I0929 17:43:05.795751 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll_19f4040e-c856-4d41-a9a0-977dbce39c98/util/0.log" Sep 29 17:43:05 crc kubenswrapper[4667]: I0929 17:43:05.824350 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll_19f4040e-c856-4d41-a9a0-977dbce39c98/pull/0.log" Sep 29 17:43:05 crc kubenswrapper[4667]: I0929 17:43:05.965203 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll_19f4040e-c856-4d41-a9a0-977dbce39c98/util/0.log" Sep 29 17:43:05 crc kubenswrapper[4667]: I0929 17:43:05.965764 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll_19f4040e-c856-4d41-a9a0-977dbce39c98/pull/0.log" Sep 29 17:43:05 crc kubenswrapper[4667]: I0929 17:43:05.976830 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_9a6e092ce660f08e14c0b0ceab3711fa43f2b70244f9df8a7a069040bcw9kll_19f4040e-c856-4d41-a9a0-977dbce39c98/extract/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.097704 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v_46414ce5-c130-445a-8c2e-a3f1e4781b9d/util/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.206574 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v_46414ce5-c130-445a-8c2e-a3f1e4781b9d/util/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.215940 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v_46414ce5-c130-445a-8c2e-a3f1e4781b9d/pull/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.216467 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v_46414ce5-c130-445a-8c2e-a3f1e4781b9d/pull/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.369427 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v_46414ce5-c130-445a-8c2e-a3f1e4781b9d/pull/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.373455 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v_46414ce5-c130-445a-8c2e-a3f1e4781b9d/extract/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.379762 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2ddpf8v_46414ce5-c130-445a-8c2e-a3f1e4781b9d/util/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.493277 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb_7f8a4805-c531-4623-a914-f76593c170ae/util/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.659276 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb_7f8a4805-c531-4623-a914-f76593c170ae/pull/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.663474 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb_7f8a4805-c531-4623-a914-f76593c170ae/pull/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.666110 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb_7f8a4805-c531-4623-a914-f76593c170ae/util/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.785257 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb_7f8a4805-c531-4623-a914-f76593c170ae/util/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.796805 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb_7f8a4805-c531-4623-a914-f76593c170ae/pull/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.807306 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_c03590272772b1d93899b6ceaa83703cf46dc8f83faf0e965a036060c0zhggb_7f8a4805-c531-4623-a914-f76593c170ae/extract/0.log" Sep 29 17:43:06 crc kubenswrapper[4667]: I0929 17:43:06.914743 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5zzk_b0a801d2-773e-4acd-905d-25dd9e9e2dcd/extract-utilities/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.061175 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5zzk_b0a801d2-773e-4acd-905d-25dd9e9e2dcd/extract-content/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.061621 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5zzk_b0a801d2-773e-4acd-905d-25dd9e9e2dcd/extract-utilities/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.088585 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5zzk_b0a801d2-773e-4acd-905d-25dd9e9e2dcd/extract-content/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.208157 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5zzk_b0a801d2-773e-4acd-905d-25dd9e9e2dcd/extract-content/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.234259 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5zzk_b0a801d2-773e-4acd-905d-25dd9e9e2dcd/extract-utilities/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.399922 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tcltq_50fa2887-a6dd-4b50-bd32-cc65c5a380ce/extract-utilities/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.461636 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-g5zzk_b0a801d2-773e-4acd-905d-25dd9e9e2dcd/registry-server/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.517141 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tcltq_50fa2887-a6dd-4b50-bd32-cc65c5a380ce/extract-utilities/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.527346 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tcltq_50fa2887-a6dd-4b50-bd32-cc65c5a380ce/extract-content/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.595104 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tcltq_50fa2887-a6dd-4b50-bd32-cc65c5a380ce/extract-content/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.711747 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tcltq_50fa2887-a6dd-4b50-bd32-cc65c5a380ce/extract-content/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.724829 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tcltq_50fa2887-a6dd-4b50-bd32-cc65c5a380ce/extract-utilities/0.log" Sep 29 17:43:07 crc kubenswrapper[4667]: I0929 17:43:07.900083 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tcltq_50fa2887-a6dd-4b50-bd32-cc65c5a380ce/registry-server/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.026459 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs_f334231f-deeb-406a-930f-53c1fb6d7b86/util/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.171128 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs_f334231f-deeb-406a-930f-53c1fb6d7b86/util/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.184070 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs_f334231f-deeb-406a-930f-53c1fb6d7b86/pull/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.185504 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs_f334231f-deeb-406a-930f-53c1fb6d7b86/pull/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.315213 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs_f334231f-deeb-406a-930f-53c1fb6d7b86/pull/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.321646 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs_f334231f-deeb-406a-930f-53c1fb6d7b86/util/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.345154 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8d7c1038c65d2785a47a2ffcc15b07abd45421e7db92f3c296d966170cvmcs_f334231f-deeb-406a-930f-53c1fb6d7b86/extract/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.447577 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4_5a3413fd-e8f4-40d4-8919-87e18ffeef7d/util/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.561514 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4_5a3413fd-e8f4-40d4-8919-87e18ffeef7d/util/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.561529 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4_5a3413fd-e8f4-40d4-8919-87e18ffeef7d/pull/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.589101 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4_5a3413fd-e8f4-40d4-8919-87e18ffeef7d/pull/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.731332 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4_5a3413fd-e8f4-40d4-8919-87e18ffeef7d/util/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.740081 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qzqrj_f8a1767f-4534-4fa7-b126-e9c0ba822350/marketplace-operator/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.744061 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4_5a3413fd-e8f4-40d4-8919-87e18ffeef7d/pull/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.744572 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_f29efc416ca216184f30dbb4b19e0f463bdcecc8ef634322abbad88d96k7cb4_5a3413fd-e8f4-40d4-8919-87e18ffeef7d/extract/0.log" Sep 29 17:43:08 crc kubenswrapper[4667]: I0929 17:43:08.899565 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hhnh5_fcdb1214-877b-4b40-9e20-749d9d50c808/extract-utilities/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.029399 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hhnh5_fcdb1214-877b-4b40-9e20-749d9d50c808/extract-content/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.030014 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hhnh5_fcdb1214-877b-4b40-9e20-749d9d50c808/extract-utilities/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.034582 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hhnh5_fcdb1214-877b-4b40-9e20-749d9d50c808/extract-content/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.197344 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hhnh5_fcdb1214-877b-4b40-9e20-749d9d50c808/extract-utilities/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.200443 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hhnh5_fcdb1214-877b-4b40-9e20-749d9d50c808/extract-content/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.222011 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7jzk2_5877d4d8-3fbf-448c-963f-c6e493fb1c64/extract-utilities/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.250083 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hhnh5_fcdb1214-877b-4b40-9e20-749d9d50c808/registry-server/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.369579 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7jzk2_5877d4d8-3fbf-448c-963f-c6e493fb1c64/extract-content/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.371983 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7jzk2_5877d4d8-3fbf-448c-963f-c6e493fb1c64/extract-utilities/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.406041 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7jzk2_5877d4d8-3fbf-448c-963f-c6e493fb1c64/extract-content/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.542371 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7jzk2_5877d4d8-3fbf-448c-963f-c6e493fb1c64/extract-utilities/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.542408 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7jzk2_5877d4d8-3fbf-448c-963f-c6e493fb1c64/extract-content/0.log" Sep 29 17:43:09 crc kubenswrapper[4667]: I0929 17:43:09.829280 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-7jzk2_5877d4d8-3fbf-448c-963f-c6e493fb1c64/registry-server/0.log" Sep 29 17:43:18 crc kubenswrapper[4667]: I0929 17:43:18.770414 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-x5wg4_04f22be1-6686-49ab-8abd-4d25428e1a8d/prometheus-operator/0.log" Sep 29 17:43:18 crc kubenswrapper[4667]: I0929 17:43:18.838190 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-l2t9q_3a9a6102-73a2-44c8-b7fe-34e0bc6f6197/prometheus-operator-admission-webhook/0.log" Sep 29 17:43:18 crc kubenswrapper[4667]: I0929 17:43:18.952666 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-646f8974b4-r79h4_677f2569-3665-45d9-bade-1aaf037a4cfb/prometheus-operator-admission-webhook/0.log" Sep 29 17:43:18 crc kubenswrapper[4667]: I0929 17:43:18.973734 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-4mj5j_e01c6c8e-211e-4afd-8125-46eec155afc3/operator/0.log" Sep 29 17:43:19 crc kubenswrapper[4667]: I0929 17:43:19.094142 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-rphbr_f926aaae-526c-4f52-99b3-4b799961c9d2/perses-operator/0.log" Sep 29 17:43:28 crc kubenswrapper[4667]: I0929 17:43:28.083981 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-679b68c759-tgpzd_fbde58f9-2869-40cd-bef8-a28336337936/kube-rbac-proxy/0.log" Sep 29 17:43:28 crc kubenswrapper[4667]: I0929 17:43:28.130196 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-679b68c759-tgpzd_fbde58f9-2869-40cd-bef8-a28336337936/manager/0.log" Sep 29 17:43:37 crc kubenswrapper[4667]: E0929 17:43:37.822446 4667 log.go:32] "PullImage from image service failed" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:43:37 crc kubenswrapper[4667]: E0929 17:43:37.823101 4667 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" image="38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822" Sep 29 17:43:37 crc kubenswrapper[4667]: E0929 17:43:37.823234 4667 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmkl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-operator-index-6b7qn_openstack-operators(7aa88991-5ef2-4166-8d37-8ce5fb207d26): ErrImagePull: rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \"http://38.129.56.34:5001/v2/\": dial tcp 38.129.56.34:5001: i/o timeout" logger="UnhandledError" Sep 29 17:43:37 crc kubenswrapper[4667]: E0929 17:43:37.824620 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = DeadlineExceeded desc = initializing source docker://38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822: pinging container registry 38.129.56.34:5001: Get \\\"http://38.129.56.34:5001/v2/\\\": dial tcp 38.129.56.34:5001: i/o timeout\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:43:39 crc kubenswrapper[4667]: I0929 17:43:39.925141 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2wmvx"] Sep 29 17:43:39 crc kubenswrapper[4667]: I0929 17:43:39.926663 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:39 crc kubenswrapper[4667]: I0929 17:43:39.939563 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2wmvx"] Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.002922 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcprw\" (UniqueName: \"kubernetes.io/projected/60773e3c-669b-446d-844e-20884ee16d35-kube-api-access-wcprw\") pod \"community-operators-2wmvx\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.003024 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-utilities\") pod \"community-operators-2wmvx\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.003055 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-catalog-content\") pod \"community-operators-2wmvx\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.105468 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcprw\" (UniqueName: \"kubernetes.io/projected/60773e3c-669b-446d-844e-20884ee16d35-kube-api-access-wcprw\") pod \"community-operators-2wmvx\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.106625 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-utilities\") pod \"community-operators-2wmvx\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.106707 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-catalog-content\") pod \"community-operators-2wmvx\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.107097 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-utilities\") pod \"community-operators-2wmvx\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.107184 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-catalog-content\") pod \"community-operators-2wmvx\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.135726 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcprw\" (UniqueName: \"kubernetes.io/projected/60773e3c-669b-446d-844e-20884ee16d35-kube-api-access-wcprw\") pod \"community-operators-2wmvx\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.244932 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:40 crc kubenswrapper[4667]: I0929 17:43:40.715454 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2wmvx"] Sep 29 17:43:41 crc kubenswrapper[4667]: I0929 17:43:41.065125 4667 generic.go:334] "Generic (PLEG): container finished" podID="60773e3c-669b-446d-844e-20884ee16d35" containerID="ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57" exitCode=0 Sep 29 17:43:41 crc kubenswrapper[4667]: I0929 17:43:41.065169 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wmvx" event={"ID":"60773e3c-669b-446d-844e-20884ee16d35","Type":"ContainerDied","Data":"ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57"} Sep 29 17:43:41 crc kubenswrapper[4667]: I0929 17:43:41.065199 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wmvx" event={"ID":"60773e3c-669b-446d-844e-20884ee16d35","Type":"ContainerStarted","Data":"07e200ea32be432d88e1b6aeec3926be59bdb9c3f2f04d393a041dbabf7aaaf1"} Sep 29 17:43:42 crc kubenswrapper[4667]: I0929 17:43:42.073887 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wmvx" event={"ID":"60773e3c-669b-446d-844e-20884ee16d35","Type":"ContainerStarted","Data":"51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc"} Sep 29 17:43:43 crc kubenswrapper[4667]: I0929 17:43:43.081340 4667 generic.go:334] "Generic (PLEG): container finished" podID="60773e3c-669b-446d-844e-20884ee16d35" containerID="51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc" exitCode=0 Sep 29 17:43:43 crc kubenswrapper[4667]: I0929 17:43:43.083165 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wmvx" event={"ID":"60773e3c-669b-446d-844e-20884ee16d35","Type":"ContainerDied","Data":"51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc"} Sep 29 17:43:44 crc kubenswrapper[4667]: I0929 17:43:44.093049 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wmvx" event={"ID":"60773e3c-669b-446d-844e-20884ee16d35","Type":"ContainerStarted","Data":"783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3"} Sep 29 17:43:44 crc kubenswrapper[4667]: I0929 17:43:44.115536 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2wmvx" podStartSLOduration=2.462798171 podStartE2EDuration="5.115517217s" podCreationTimestamp="2025-09-29 17:43:39 +0000 UTC" firstStartedPulling="2025-09-29 17:43:41.067962138 +0000 UTC m=+2049.565808907" lastFinishedPulling="2025-09-29 17:43:43.720681184 +0000 UTC m=+2052.218527953" observedRunningTime="2025-09-29 17:43:44.108923427 +0000 UTC m=+2052.606770196" watchObservedRunningTime="2025-09-29 17:43:44.115517217 +0000 UTC m=+2052.613363986" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.314088 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z9c8t"] Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.315868 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.326882 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9c8t"] Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.501419 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-catalog-content\") pod \"redhat-marketplace-z9c8t\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.501583 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-utilities\") pod \"redhat-marketplace-z9c8t\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.501654 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7smpc\" (UniqueName: \"kubernetes.io/projected/447bd602-f83b-4dc3-a51b-685931e1dc90-kube-api-access-7smpc\") pod \"redhat-marketplace-z9c8t\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.603817 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7smpc\" (UniqueName: \"kubernetes.io/projected/447bd602-f83b-4dc3-a51b-685931e1dc90-kube-api-access-7smpc\") pod \"redhat-marketplace-z9c8t\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.603999 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-catalog-content\") pod \"redhat-marketplace-z9c8t\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.604150 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-utilities\") pod \"redhat-marketplace-z9c8t\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.604492 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-catalog-content\") pod \"redhat-marketplace-z9c8t\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.604786 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-utilities\") pod \"redhat-marketplace-z9c8t\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.635875 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7smpc\" (UniqueName: \"kubernetes.io/projected/447bd602-f83b-4dc3-a51b-685931e1dc90-kube-api-access-7smpc\") pod \"redhat-marketplace-z9c8t\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:46 crc kubenswrapper[4667]: I0929 17:43:46.645245 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:47 crc kubenswrapper[4667]: I0929 17:43:47.071954 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9c8t"] Sep 29 17:43:47 crc kubenswrapper[4667]: I0929 17:43:47.133043 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9c8t" event={"ID":"447bd602-f83b-4dc3-a51b-685931e1dc90","Type":"ContainerStarted","Data":"d53ae6153fc03a1a1f3e204310b09f0e882075b95b726c50cde9c332bce4d96e"} Sep 29 17:43:48 crc kubenswrapper[4667]: I0929 17:43:48.146906 4667 generic.go:334] "Generic (PLEG): container finished" podID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerID="c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5" exitCode=0 Sep 29 17:43:48 crc kubenswrapper[4667]: I0929 17:43:48.147359 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9c8t" event={"ID":"447bd602-f83b-4dc3-a51b-685931e1dc90","Type":"ContainerDied","Data":"c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5"} Sep 29 17:43:49 crc kubenswrapper[4667]: I0929 17:43:49.160076 4667 generic.go:334] "Generic (PLEG): container finished" podID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerID="97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001" exitCode=0 Sep 29 17:43:49 crc kubenswrapper[4667]: I0929 17:43:49.160160 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9c8t" event={"ID":"447bd602-f83b-4dc3-a51b-685931e1dc90","Type":"ContainerDied","Data":"97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001"} Sep 29 17:43:49 crc kubenswrapper[4667]: E0929 17:43:49.817262 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:43:50 crc kubenswrapper[4667]: I0929 17:43:50.172533 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9c8t" event={"ID":"447bd602-f83b-4dc3-a51b-685931e1dc90","Type":"ContainerStarted","Data":"08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb"} Sep 29 17:43:50 crc kubenswrapper[4667]: I0929 17:43:50.191480 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z9c8t" podStartSLOduration=2.625860218 podStartE2EDuration="4.191460753s" podCreationTimestamp="2025-09-29 17:43:46 +0000 UTC" firstStartedPulling="2025-09-29 17:43:48.158922236 +0000 UTC m=+2056.656769005" lastFinishedPulling="2025-09-29 17:43:49.724522771 +0000 UTC m=+2058.222369540" observedRunningTime="2025-09-29 17:43:50.185127414 +0000 UTC m=+2058.682974183" watchObservedRunningTime="2025-09-29 17:43:50.191460753 +0000 UTC m=+2058.689307513" Sep 29 17:43:50 crc kubenswrapper[4667]: I0929 17:43:50.245584 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:50 crc kubenswrapper[4667]: I0929 17:43:50.245894 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:50 crc kubenswrapper[4667]: I0929 17:43:50.286459 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:51 crc kubenswrapper[4667]: I0929 17:43:51.223088 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:52 crc kubenswrapper[4667]: I0929 17:43:52.902129 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2wmvx"] Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.197271 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2wmvx" podUID="60773e3c-669b-446d-844e-20884ee16d35" containerName="registry-server" containerID="cri-o://783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3" gracePeriod=2 Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.572943 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.743484 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-catalog-content\") pod \"60773e3c-669b-446d-844e-20884ee16d35\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.743776 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-utilities\") pod \"60773e3c-669b-446d-844e-20884ee16d35\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.743907 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcprw\" (UniqueName: \"kubernetes.io/projected/60773e3c-669b-446d-844e-20884ee16d35-kube-api-access-wcprw\") pod \"60773e3c-669b-446d-844e-20884ee16d35\" (UID: \"60773e3c-669b-446d-844e-20884ee16d35\") " Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.746069 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-utilities" (OuterVolumeSpecName: "utilities") pod "60773e3c-669b-446d-844e-20884ee16d35" (UID: "60773e3c-669b-446d-844e-20884ee16d35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.758957 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60773e3c-669b-446d-844e-20884ee16d35-kube-api-access-wcprw" (OuterVolumeSpecName: "kube-api-access-wcprw") pod "60773e3c-669b-446d-844e-20884ee16d35" (UID: "60773e3c-669b-446d-844e-20884ee16d35"). InnerVolumeSpecName "kube-api-access-wcprw". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.778175 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60773e3c-669b-446d-844e-20884ee16d35" (UID: "60773e3c-669b-446d-844e-20884ee16d35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.846758 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.846783 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60773e3c-669b-446d-844e-20884ee16d35-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:43:54 crc kubenswrapper[4667]: I0929 17:43:54.846794 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcprw\" (UniqueName: \"kubernetes.io/projected/60773e3c-669b-446d-844e-20884ee16d35-kube-api-access-wcprw\") on node \"crc\" DevicePath \"\"" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.205875 4667 generic.go:334] "Generic (PLEG): container finished" podID="60773e3c-669b-446d-844e-20884ee16d35" containerID="783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3" exitCode=0 Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.205959 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wmvx" event={"ID":"60773e3c-669b-446d-844e-20884ee16d35","Type":"ContainerDied","Data":"783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3"} Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.205995 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2wmvx" event={"ID":"60773e3c-669b-446d-844e-20884ee16d35","Type":"ContainerDied","Data":"07e200ea32be432d88e1b6aeec3926be59bdb9c3f2f04d393a041dbabf7aaaf1"} Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.206012 4667 scope.go:117] "RemoveContainer" containerID="783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.206678 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2wmvx" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.233819 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2wmvx"] Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.245707 4667 scope.go:117] "RemoveContainer" containerID="51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.250028 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2wmvx"] Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.266197 4667 scope.go:117] "RemoveContainer" containerID="ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.285037 4667 scope.go:117] "RemoveContainer" containerID="783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3" Sep 29 17:43:55 crc kubenswrapper[4667]: E0929 17:43:55.286027 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3\": container with ID starting with 783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3 not found: ID does not exist" containerID="783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.286051 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3"} err="failed to get container status \"783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3\": rpc error: code = NotFound desc = could not find container \"783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3\": container with ID starting with 783441090f66a1902a95397fea2b7d5b81b01d49429e4027aec81804aa84efa3 not found: ID does not exist" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.286070 4667 scope.go:117] "RemoveContainer" containerID="51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc" Sep 29 17:43:55 crc kubenswrapper[4667]: E0929 17:43:55.286372 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc\": container with ID starting with 51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc not found: ID does not exist" containerID="51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.286393 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc"} err="failed to get container status \"51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc\": rpc error: code = NotFound desc = could not find container \"51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc\": container with ID starting with 51a9d5ecfb60409387a7858e76ee00aa388a00dbe52a76e73f78221a72f0b6dc not found: ID does not exist" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.286412 4667 scope.go:117] "RemoveContainer" containerID="ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57" Sep 29 17:43:55 crc kubenswrapper[4667]: E0929 17:43:55.286705 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57\": container with ID starting with ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57 not found: ID does not exist" containerID="ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.286720 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57"} err="failed to get container status \"ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57\": rpc error: code = NotFound desc = could not find container \"ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57\": container with ID starting with ec1892af7eda7b8ff7bdd635d4be6279150278f156fdf10ec36b1473f93e5e57 not found: ID does not exist" Sep 29 17:43:55 crc kubenswrapper[4667]: I0929 17:43:55.834343 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60773e3c-669b-446d-844e-20884ee16d35" path="/var/lib/kubelet/pods/60773e3c-669b-446d-844e-20884ee16d35/volumes" Sep 29 17:43:56 crc kubenswrapper[4667]: I0929 17:43:56.646344 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:56 crc kubenswrapper[4667]: I0929 17:43:56.646397 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:56 crc kubenswrapper[4667]: I0929 17:43:56.684314 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:57 crc kubenswrapper[4667]: I0929 17:43:57.255529 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:58 crc kubenswrapper[4667]: I0929 17:43:58.304736 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9c8t"] Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.230017 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z9c8t" podUID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerName="registry-server" containerID="cri-o://08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb" gracePeriod=2 Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.597228 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.724129 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7smpc\" (UniqueName: \"kubernetes.io/projected/447bd602-f83b-4dc3-a51b-685931e1dc90-kube-api-access-7smpc\") pod \"447bd602-f83b-4dc3-a51b-685931e1dc90\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.726738 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-utilities\") pod \"447bd602-f83b-4dc3-a51b-685931e1dc90\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.728245 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-catalog-content\") pod \"447bd602-f83b-4dc3-a51b-685931e1dc90\" (UID: \"447bd602-f83b-4dc3-a51b-685931e1dc90\") " Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.728104 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-utilities" (OuterVolumeSpecName: "utilities") pod "447bd602-f83b-4dc3-a51b-685931e1dc90" (UID: "447bd602-f83b-4dc3-a51b-685931e1dc90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.739392 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "447bd602-f83b-4dc3-a51b-685931e1dc90" (UID: "447bd602-f83b-4dc3-a51b-685931e1dc90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.739659 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/447bd602-f83b-4dc3-a51b-685931e1dc90-kube-api-access-7smpc" (OuterVolumeSpecName: "kube-api-access-7smpc") pod "447bd602-f83b-4dc3-a51b-685931e1dc90" (UID: "447bd602-f83b-4dc3-a51b-685931e1dc90"). InnerVolumeSpecName "kube-api-access-7smpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.739731 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.739758 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/447bd602-f83b-4dc3-a51b-685931e1dc90-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:43:59 crc kubenswrapper[4667]: I0929 17:43:59.841538 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7smpc\" (UniqueName: \"kubernetes.io/projected/447bd602-f83b-4dc3-a51b-685931e1dc90-kube-api-access-7smpc\") on node \"crc\" DevicePath \"\"" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.254061 4667 generic.go:334] "Generic (PLEG): container finished" podID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerID="08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb" exitCode=0 Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.254126 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9c8t" event={"ID":"447bd602-f83b-4dc3-a51b-685931e1dc90","Type":"ContainerDied","Data":"08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb"} Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.254476 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9c8t" event={"ID":"447bd602-f83b-4dc3-a51b-685931e1dc90","Type":"ContainerDied","Data":"d53ae6153fc03a1a1f3e204310b09f0e882075b95b726c50cde9c332bce4d96e"} Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.254503 4667 scope.go:117] "RemoveContainer" containerID="08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.254187 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9c8t" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.276163 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9c8t"] Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.281592 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9c8t"] Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.283724 4667 scope.go:117] "RemoveContainer" containerID="97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.302467 4667 scope.go:117] "RemoveContainer" containerID="c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.317967 4667 scope.go:117] "RemoveContainer" containerID="08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb" Sep 29 17:44:00 crc kubenswrapper[4667]: E0929 17:44:00.318230 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb\": container with ID starting with 08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb not found: ID does not exist" containerID="08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.318262 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb"} err="failed to get container status \"08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb\": rpc error: code = NotFound desc = could not find container \"08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb\": container with ID starting with 08a304b956a9b053807325a300a36aa9332a71bc1bba64072085d5db83688ecb not found: ID does not exist" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.318282 4667 scope.go:117] "RemoveContainer" containerID="97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001" Sep 29 17:44:00 crc kubenswrapper[4667]: E0929 17:44:00.318521 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001\": container with ID starting with 97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001 not found: ID does not exist" containerID="97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.318568 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001"} err="failed to get container status \"97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001\": rpc error: code = NotFound desc = could not find container \"97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001\": container with ID starting with 97f005b9f4848e751aae8cd3eea7fbb9308713f9c652cb2cab419bb9a59fd001 not found: ID does not exist" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.318587 4667 scope.go:117] "RemoveContainer" containerID="c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5" Sep 29 17:44:00 crc kubenswrapper[4667]: E0929 17:44:00.318813 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5\": container with ID starting with c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5 not found: ID does not exist" containerID="c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.318838 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5"} err="failed to get container status \"c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5\": rpc error: code = NotFound desc = could not find container \"c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5\": container with ID starting with c856e6d114daf79084885847dac1d18786f9560b052a4e63d0e60b415765ada5 not found: ID does not exist" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.509818 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l5tcl"] Sep 29 17:44:00 crc kubenswrapper[4667]: E0929 17:44:00.510128 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerName="registry-server" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.510141 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerName="registry-server" Sep 29 17:44:00 crc kubenswrapper[4667]: E0929 17:44:00.510152 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60773e3c-669b-446d-844e-20884ee16d35" containerName="registry-server" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.510158 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="60773e3c-669b-446d-844e-20884ee16d35" containerName="registry-server" Sep 29 17:44:00 crc kubenswrapper[4667]: E0929 17:44:00.510176 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60773e3c-669b-446d-844e-20884ee16d35" containerName="extract-content" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.510181 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="60773e3c-669b-446d-844e-20884ee16d35" containerName="extract-content" Sep 29 17:44:00 crc kubenswrapper[4667]: E0929 17:44:00.510189 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerName="extract-utilities" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.510194 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerName="extract-utilities" Sep 29 17:44:00 crc kubenswrapper[4667]: E0929 17:44:00.510208 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60773e3c-669b-446d-844e-20884ee16d35" containerName="extract-utilities" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.510213 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="60773e3c-669b-446d-844e-20884ee16d35" containerName="extract-utilities" Sep 29 17:44:00 crc kubenswrapper[4667]: E0929 17:44:00.510223 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerName="extract-content" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.510230 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerName="extract-content" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.510389 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="60773e3c-669b-446d-844e-20884ee16d35" containerName="registry-server" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.510402 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="447bd602-f83b-4dc3-a51b-685931e1dc90" containerName="registry-server" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.511401 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.521930 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5tcl"] Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.554727 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrvm2\" (UniqueName: \"kubernetes.io/projected/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-kube-api-access-wrvm2\") pod \"redhat-operators-l5tcl\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.555144 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-catalog-content\") pod \"redhat-operators-l5tcl\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.555342 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-utilities\") pod \"redhat-operators-l5tcl\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.657902 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrvm2\" (UniqueName: \"kubernetes.io/projected/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-kube-api-access-wrvm2\") pod \"redhat-operators-l5tcl\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.659093 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-catalog-content\") pod \"redhat-operators-l5tcl\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.659260 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-utilities\") pod \"redhat-operators-l5tcl\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.659622 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-utilities\") pod \"redhat-operators-l5tcl\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.659619 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-catalog-content\") pod \"redhat-operators-l5tcl\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.682650 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrvm2\" (UniqueName: \"kubernetes.io/projected/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-kube-api-access-wrvm2\") pod \"redhat-operators-l5tcl\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:00 crc kubenswrapper[4667]: I0929 17:44:00.829677 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:01 crc kubenswrapper[4667]: I0929 17:44:01.225547 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5tcl"] Sep 29 17:44:01 crc kubenswrapper[4667]: I0929 17:44:01.279513 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5tcl" event={"ID":"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0","Type":"ContainerStarted","Data":"023a0f28207c546c9d808c7d358db46ba607be4a6db3f925e30eec19b447dc9e"} Sep 29 17:44:01 crc kubenswrapper[4667]: I0929 17:44:01.822496 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="447bd602-f83b-4dc3-a51b-685931e1dc90" path="/var/lib/kubelet/pods/447bd602-f83b-4dc3-a51b-685931e1dc90/volumes" Sep 29 17:44:02 crc kubenswrapper[4667]: I0929 17:44:02.285829 4667 generic.go:334] "Generic (PLEG): container finished" podID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerID="7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e" exitCode=0 Sep 29 17:44:02 crc kubenswrapper[4667]: I0929 17:44:02.285874 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5tcl" event={"ID":"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0","Type":"ContainerDied","Data":"7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e"} Sep 29 17:44:02 crc kubenswrapper[4667]: I0929 17:44:02.923386 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7jt6b"] Sep 29 17:44:02 crc kubenswrapper[4667]: I0929 17:44:02.926358 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:02 crc kubenswrapper[4667]: I0929 17:44:02.932325 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jt6b"] Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.008677 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-utilities\") pod \"certified-operators-7jt6b\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.008739 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fprkx\" (UniqueName: \"kubernetes.io/projected/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-kube-api-access-fprkx\") pod \"certified-operators-7jt6b\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.008907 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-catalog-content\") pod \"certified-operators-7jt6b\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.110765 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-catalog-content\") pod \"certified-operators-7jt6b\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.111242 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-catalog-content\") pod \"certified-operators-7jt6b\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.111803 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-utilities\") pod \"certified-operators-7jt6b\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.111926 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fprkx\" (UniqueName: \"kubernetes.io/projected/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-kube-api-access-fprkx\") pod \"certified-operators-7jt6b\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.112113 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-utilities\") pod \"certified-operators-7jt6b\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.151393 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fprkx\" (UniqueName: \"kubernetes.io/projected/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-kube-api-access-fprkx\") pod \"certified-operators-7jt6b\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.253546 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:03 crc kubenswrapper[4667]: I0929 17:44:03.718346 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7jt6b"] Sep 29 17:44:03 crc kubenswrapper[4667]: E0929 17:44:03.821134 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:44:04 crc kubenswrapper[4667]: I0929 17:44:04.310950 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5tcl" event={"ID":"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0","Type":"ContainerStarted","Data":"a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4"} Sep 29 17:44:04 crc kubenswrapper[4667]: I0929 17:44:04.314473 4667 generic.go:334] "Generic (PLEG): container finished" podID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerID="ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9" exitCode=0 Sep 29 17:44:04 crc kubenswrapper[4667]: I0929 17:44:04.314508 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jt6b" event={"ID":"c62a990b-3b5d-4a6e-b415-58c3f7cf5178","Type":"ContainerDied","Data":"ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9"} Sep 29 17:44:04 crc kubenswrapper[4667]: I0929 17:44:04.314549 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jt6b" event={"ID":"c62a990b-3b5d-4a6e-b415-58c3f7cf5178","Type":"ContainerStarted","Data":"c003e371b0602d7285137537c0046b8d32b942c61c3b817a21702d9ee5e94a8f"} Sep 29 17:44:05 crc kubenswrapper[4667]: I0929 17:44:05.323554 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jt6b" event={"ID":"c62a990b-3b5d-4a6e-b415-58c3f7cf5178","Type":"ContainerStarted","Data":"913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc"} Sep 29 17:44:05 crc kubenswrapper[4667]: I0929 17:44:05.327502 4667 generic.go:334] "Generic (PLEG): container finished" podID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerID="a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4" exitCode=0 Sep 29 17:44:05 crc kubenswrapper[4667]: I0929 17:44:05.327545 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5tcl" event={"ID":"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0","Type":"ContainerDied","Data":"a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4"} Sep 29 17:44:06 crc kubenswrapper[4667]: I0929 17:44:06.349809 4667 generic.go:334] "Generic (PLEG): container finished" podID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerID="913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc" exitCode=0 Sep 29 17:44:06 crc kubenswrapper[4667]: I0929 17:44:06.350385 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jt6b" event={"ID":"c62a990b-3b5d-4a6e-b415-58c3f7cf5178","Type":"ContainerDied","Data":"913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc"} Sep 29 17:44:06 crc kubenswrapper[4667]: I0929 17:44:06.352677 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5tcl" event={"ID":"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0","Type":"ContainerStarted","Data":"91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd"} Sep 29 17:44:07 crc kubenswrapper[4667]: I0929 17:44:07.360782 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jt6b" event={"ID":"c62a990b-3b5d-4a6e-b415-58c3f7cf5178","Type":"ContainerStarted","Data":"eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2"} Sep 29 17:44:07 crc kubenswrapper[4667]: I0929 17:44:07.362741 4667 generic.go:334] "Generic (PLEG): container finished" podID="549699f7-0860-44fd-9bce-a37ce789396c" containerID="e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54" exitCode=0 Sep 29 17:44:07 crc kubenswrapper[4667]: I0929 17:44:07.363368 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bns76/must-gather-m6926" event={"ID":"549699f7-0860-44fd-9bce-a37ce789396c","Type":"ContainerDied","Data":"e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54"} Sep 29 17:44:07 crc kubenswrapper[4667]: I0929 17:44:07.363683 4667 scope.go:117] "RemoveContainer" containerID="e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54" Sep 29 17:44:07 crc kubenswrapper[4667]: I0929 17:44:07.382174 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7jt6b" podStartSLOduration=2.8552970269999998 podStartE2EDuration="5.382157096s" podCreationTimestamp="2025-09-29 17:44:02 +0000 UTC" firstStartedPulling="2025-09-29 17:44:04.320404314 +0000 UTC m=+2072.818251093" lastFinishedPulling="2025-09-29 17:44:06.847264392 +0000 UTC m=+2075.345111162" observedRunningTime="2025-09-29 17:44:07.379408716 +0000 UTC m=+2075.877255485" watchObservedRunningTime="2025-09-29 17:44:07.382157096 +0000 UTC m=+2075.880003855" Sep 29 17:44:07 crc kubenswrapper[4667]: I0929 17:44:07.383466 4667 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l5tcl" podStartSLOduration=3.8084047869999997 podStartE2EDuration="7.383457237s" podCreationTimestamp="2025-09-29 17:44:00 +0000 UTC" firstStartedPulling="2025-09-29 17:44:02.288227729 +0000 UTC m=+2070.786074489" lastFinishedPulling="2025-09-29 17:44:05.863280169 +0000 UTC m=+2074.361126939" observedRunningTime="2025-09-29 17:44:06.384728129 +0000 UTC m=+2074.882574898" watchObservedRunningTime="2025-09-29 17:44:07.383457237 +0000 UTC m=+2075.881304006" Sep 29 17:44:07 crc kubenswrapper[4667]: I0929 17:44:07.813021 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bns76_must-gather-m6926_549699f7-0860-44fd-9bce-a37ce789396c/gather/0.log" Sep 29 17:44:10 crc kubenswrapper[4667]: I0929 17:44:10.830736 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:10 crc kubenswrapper[4667]: I0929 17:44:10.831432 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:10 crc kubenswrapper[4667]: I0929 17:44:10.871376 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:11 crc kubenswrapper[4667]: I0929 17:44:11.420828 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:12 crc kubenswrapper[4667]: I0929 17:44:12.707027 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5tcl"] Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.254414 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.254468 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.287912 4667 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.404883 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l5tcl" podUID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerName="registry-server" containerID="cri-o://91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd" gracePeriod=2 Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.440091 4667 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.780867 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.802377 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrvm2\" (UniqueName: \"kubernetes.io/projected/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-kube-api-access-wrvm2\") pod \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.802539 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-utilities\") pod \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.802656 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-catalog-content\") pod \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\" (UID: \"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0\") " Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.803183 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-utilities" (OuterVolumeSpecName: "utilities") pod "d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" (UID: "d26c879c-d92c-4525-9cd7-9a2c09b2c9b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.808673 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-kube-api-access-wrvm2" (OuterVolumeSpecName: "kube-api-access-wrvm2") pod "d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" (UID: "d26c879c-d92c-4525-9cd7-9a2c09b2c9b0"). InnerVolumeSpecName "kube-api-access-wrvm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.869694 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" (UID: "d26c879c-d92c-4525-9cd7-9a2c09b2c9b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.904224 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrvm2\" (UniqueName: \"kubernetes.io/projected/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-kube-api-access-wrvm2\") on node \"crc\" DevicePath \"\"" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.904252 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:44:13 crc kubenswrapper[4667]: I0929 17:44:13.904263 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.063479 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bns76/must-gather-m6926"] Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.063718 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-bns76/must-gather-m6926" podUID="549699f7-0860-44fd-9bce-a37ce789396c" containerName="copy" containerID="cri-o://818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6" gracePeriod=2 Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.067965 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bns76/must-gather-m6926"] Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.367372 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bns76_must-gather-m6926_549699f7-0860-44fd-9bce-a37ce789396c/copy/0.log" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.368378 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.411570 4667 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bns76_must-gather-m6926_549699f7-0860-44fd-9bce-a37ce789396c/copy/0.log" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.412167 4667 generic.go:334] "Generic (PLEG): container finished" podID="549699f7-0860-44fd-9bce-a37ce789396c" containerID="818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6" exitCode=143 Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.412215 4667 scope.go:117] "RemoveContainer" containerID="818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.412253 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bns76/must-gather-m6926" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.412666 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/549699f7-0860-44fd-9bce-a37ce789396c-must-gather-output\") pod \"549699f7-0860-44fd-9bce-a37ce789396c\" (UID: \"549699f7-0860-44fd-9bce-a37ce789396c\") " Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.412731 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsm4f\" (UniqueName: \"kubernetes.io/projected/549699f7-0860-44fd-9bce-a37ce789396c-kube-api-access-fsm4f\") pod \"549699f7-0860-44fd-9bce-a37ce789396c\" (UID: \"549699f7-0860-44fd-9bce-a37ce789396c\") " Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.414675 4667 generic.go:334] "Generic (PLEG): container finished" podID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerID="91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd" exitCode=0 Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.415450 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5tcl" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.415768 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5tcl" event={"ID":"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0","Type":"ContainerDied","Data":"91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd"} Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.415807 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5tcl" event={"ID":"d26c879c-d92c-4525-9cd7-9a2c09b2c9b0","Type":"ContainerDied","Data":"023a0f28207c546c9d808c7d358db46ba607be4a6db3f925e30eec19b447dc9e"} Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.416226 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/549699f7-0860-44fd-9bce-a37ce789396c-kube-api-access-fsm4f" (OuterVolumeSpecName: "kube-api-access-fsm4f") pod "549699f7-0860-44fd-9bce-a37ce789396c" (UID: "549699f7-0860-44fd-9bce-a37ce789396c"). InnerVolumeSpecName "kube-api-access-fsm4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.449880 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5tcl"] Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.454269 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l5tcl"] Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.455925 4667 scope.go:117] "RemoveContainer" containerID="e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.474824 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/549699f7-0860-44fd-9bce-a37ce789396c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "549699f7-0860-44fd-9bce-a37ce789396c" (UID: "549699f7-0860-44fd-9bce-a37ce789396c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.488327 4667 scope.go:117] "RemoveContainer" containerID="818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6" Sep 29 17:44:14 crc kubenswrapper[4667]: E0929 17:44:14.488668 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6\": container with ID starting with 818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6 not found: ID does not exist" containerID="818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.488704 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6"} err="failed to get container status \"818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6\": rpc error: code = NotFound desc = could not find container \"818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6\": container with ID starting with 818b79430401399f31303390fb955da408da5945fdc7fd6849cb4bfcffc2dff6 not found: ID does not exist" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.488726 4667 scope.go:117] "RemoveContainer" containerID="e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54" Sep 29 17:44:14 crc kubenswrapper[4667]: E0929 17:44:14.488997 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54\": container with ID starting with e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54 not found: ID does not exist" containerID="e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.489019 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54"} err="failed to get container status \"e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54\": rpc error: code = NotFound desc = could not find container \"e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54\": container with ID starting with e2e13230b9a953d8198c5422f29393ec6caa661974e037bf3db5b6fa5dd88d54 not found: ID does not exist" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.489035 4667 scope.go:117] "RemoveContainer" containerID="91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.502822 4667 scope.go:117] "RemoveContainer" containerID="a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.514774 4667 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/549699f7-0860-44fd-9bce-a37ce789396c-must-gather-output\") on node \"crc\" DevicePath \"\"" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.514795 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsm4f\" (UniqueName: \"kubernetes.io/projected/549699f7-0860-44fd-9bce-a37ce789396c-kube-api-access-fsm4f\") on node \"crc\" DevicePath \"\"" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.515910 4667 scope.go:117] "RemoveContainer" containerID="7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.544948 4667 scope.go:117] "RemoveContainer" containerID="91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd" Sep 29 17:44:14 crc kubenswrapper[4667]: E0929 17:44:14.545334 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd\": container with ID starting with 91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd not found: ID does not exist" containerID="91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.545372 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd"} err="failed to get container status \"91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd\": rpc error: code = NotFound desc = could not find container \"91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd\": container with ID starting with 91339669f7f62fb25746efc603da2af4f2ef50a785739d588db276708afaf0bd not found: ID does not exist" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.545392 4667 scope.go:117] "RemoveContainer" containerID="a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4" Sep 29 17:44:14 crc kubenswrapper[4667]: E0929 17:44:14.545937 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4\": container with ID starting with a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4 not found: ID does not exist" containerID="a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.545983 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4"} err="failed to get container status \"a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4\": rpc error: code = NotFound desc = could not find container \"a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4\": container with ID starting with a2c94509e2eb230db80c9d36c1d2d3df3cb5a94390ac71897af7871579cd04d4 not found: ID does not exist" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.546017 4667 scope.go:117] "RemoveContainer" containerID="7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e" Sep 29 17:44:14 crc kubenswrapper[4667]: E0929 17:44:14.547405 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e\": container with ID starting with 7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e not found: ID does not exist" containerID="7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e" Sep 29 17:44:14 crc kubenswrapper[4667]: I0929 17:44:14.547478 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e"} err="failed to get container status \"7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e\": rpc error: code = NotFound desc = could not find container \"7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e\": container with ID starting with 7499a5b27511a4833c254c9c7915ed8360a986b2a36006566d2759b8e28f714e not found: ID does not exist" Sep 29 17:44:15 crc kubenswrapper[4667]: I0929 17:44:15.702371 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jt6b"] Sep 29 17:44:15 crc kubenswrapper[4667]: I0929 17:44:15.703090 4667 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7jt6b" podUID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerName="registry-server" containerID="cri-o://eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2" gracePeriod=2 Sep 29 17:44:15 crc kubenswrapper[4667]: I0929 17:44:15.825618 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="549699f7-0860-44fd-9bce-a37ce789396c" path="/var/lib/kubelet/pods/549699f7-0860-44fd-9bce-a37ce789396c/volumes" Sep 29 17:44:15 crc kubenswrapper[4667]: I0929 17:44:15.826517 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" path="/var/lib/kubelet/pods/d26c879c-d92c-4525-9cd7-9a2c09b2c9b0/volumes" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.117485 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.139699 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-utilities\") pod \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.140322 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-utilities" (OuterVolumeSpecName: "utilities") pod "c62a990b-3b5d-4a6e-b415-58c3f7cf5178" (UID: "c62a990b-3b5d-4a6e-b415-58c3f7cf5178"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.140397 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fprkx\" (UniqueName: \"kubernetes.io/projected/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-kube-api-access-fprkx\") pod \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.141100 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-catalog-content\") pod \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\" (UID: \"c62a990b-3b5d-4a6e-b415-58c3f7cf5178\") " Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.141557 4667 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-utilities\") on node \"crc\" DevicePath \"\"" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.149291 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-kube-api-access-fprkx" (OuterVolumeSpecName: "kube-api-access-fprkx") pod "c62a990b-3b5d-4a6e-b415-58c3f7cf5178" (UID: "c62a990b-3b5d-4a6e-b415-58c3f7cf5178"). InnerVolumeSpecName "kube-api-access-fprkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.174280 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c62a990b-3b5d-4a6e-b415-58c3f7cf5178" (UID: "c62a990b-3b5d-4a6e-b415-58c3f7cf5178"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.244589 4667 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-catalog-content\") on node \"crc\" DevicePath \"\"" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.244640 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fprkx\" (UniqueName: \"kubernetes.io/projected/c62a990b-3b5d-4a6e-b415-58c3f7cf5178-kube-api-access-fprkx\") on node \"crc\" DevicePath \"\"" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.435506 4667 generic.go:334] "Generic (PLEG): container finished" podID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerID="eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2" exitCode=0 Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.435555 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jt6b" event={"ID":"c62a990b-3b5d-4a6e-b415-58c3f7cf5178","Type":"ContainerDied","Data":"eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2"} Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.435600 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7jt6b" event={"ID":"c62a990b-3b5d-4a6e-b415-58c3f7cf5178","Type":"ContainerDied","Data":"c003e371b0602d7285137537c0046b8d32b942c61c3b817a21702d9ee5e94a8f"} Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.435636 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7jt6b" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.435651 4667 scope.go:117] "RemoveContainer" containerID="eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.456621 4667 scope.go:117] "RemoveContainer" containerID="913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.468051 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7jt6b"] Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.474904 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7jt6b"] Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.492700 4667 scope.go:117] "RemoveContainer" containerID="ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.511798 4667 scope.go:117] "RemoveContainer" containerID="eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2" Sep 29 17:44:16 crc kubenswrapper[4667]: E0929 17:44:16.512223 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2\": container with ID starting with eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2 not found: ID does not exist" containerID="eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.512273 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2"} err="failed to get container status \"eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2\": rpc error: code = NotFound desc = could not find container \"eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2\": container with ID starting with eaf9e6ce0468c12a14925065624e56b0595ae5bce5939aba3b9be7fbf801cde2 not found: ID does not exist" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.512316 4667 scope.go:117] "RemoveContainer" containerID="913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc" Sep 29 17:44:16 crc kubenswrapper[4667]: E0929 17:44:16.512659 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc\": container with ID starting with 913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc not found: ID does not exist" containerID="913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.512697 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc"} err="failed to get container status \"913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc\": rpc error: code = NotFound desc = could not find container \"913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc\": container with ID starting with 913840dbae3ecd6986c10f62c54422ca5c1a304bff8cc49908ab71e5e73f13bc not found: ID does not exist" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.512717 4667 scope.go:117] "RemoveContainer" containerID="ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9" Sep 29 17:44:16 crc kubenswrapper[4667]: E0929 17:44:16.513272 4667 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9\": container with ID starting with ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9 not found: ID does not exist" containerID="ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9" Sep 29 17:44:16 crc kubenswrapper[4667]: I0929 17:44:16.513308 4667 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9"} err="failed to get container status \"ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9\": rpc error: code = NotFound desc = could not find container \"ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9\": container with ID starting with ab48e91d7d4fac4a126c82a3e178322dd714c8ed676516bde89deca883d0ebc9 not found: ID does not exist" Sep 29 17:44:17 crc kubenswrapper[4667]: E0929 17:44:17.817355 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:44:17 crc kubenswrapper[4667]: I0929 17:44:17.824897 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" path="/var/lib/kubelet/pods/c62a990b-3b5d-4a6e-b415-58c3f7cf5178/volumes" Sep 29 17:44:29 crc kubenswrapper[4667]: E0929 17:44:29.817514 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:44:42 crc kubenswrapper[4667]: E0929 17:44:42.817938 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:44:56 crc kubenswrapper[4667]: E0929 17:44:56.816985 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:44:56 crc kubenswrapper[4667]: I0929 17:44:56.897259 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:44:56 crc kubenswrapper[4667]: I0929 17:44:56.897301 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.124464 4667 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb"] Sep 29 17:45:00 crc kubenswrapper[4667]: E0929 17:45:00.124916 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549699f7-0860-44fd-9bce-a37ce789396c" containerName="gather" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.124928 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="549699f7-0860-44fd-9bce-a37ce789396c" containerName="gather" Sep 29 17:45:00 crc kubenswrapper[4667]: E0929 17:45:00.124947 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerName="extract-content" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.124953 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerName="extract-content" Sep 29 17:45:00 crc kubenswrapper[4667]: E0929 17:45:00.124962 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerName="extract-content" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.124969 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerName="extract-content" Sep 29 17:45:00 crc kubenswrapper[4667]: E0929 17:45:00.124978 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerName="extract-utilities" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.124984 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerName="extract-utilities" Sep 29 17:45:00 crc kubenswrapper[4667]: E0929 17:45:00.124992 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerName="registry-server" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.124997 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerName="registry-server" Sep 29 17:45:00 crc kubenswrapper[4667]: E0929 17:45:00.125013 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549699f7-0860-44fd-9bce-a37ce789396c" containerName="copy" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.125018 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="549699f7-0860-44fd-9bce-a37ce789396c" containerName="copy" Sep 29 17:45:00 crc kubenswrapper[4667]: E0929 17:45:00.125026 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerName="extract-utilities" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.125032 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerName="extract-utilities" Sep 29 17:45:00 crc kubenswrapper[4667]: E0929 17:45:00.125045 4667 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerName="registry-server" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.125050 4667 state_mem.go:107] "Deleted CPUSet assignment" podUID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerName="registry-server" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.125166 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="d26c879c-d92c-4525-9cd7-9a2c09b2c9b0" containerName="registry-server" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.125181 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="c62a990b-3b5d-4a6e-b415-58c3f7cf5178" containerName="registry-server" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.125191 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="549699f7-0860-44fd-9bce-a37ce789396c" containerName="copy" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.125199 4667 memory_manager.go:354] "RemoveStaleState removing state" podUID="549699f7-0860-44fd-9bce-a37ce789396c" containerName="gather" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.125651 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.130762 4667 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.131016 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb"] Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.131259 4667 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.294689 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f0797db-eac1-4318-8528-17c07458885a-secret-volume\") pod \"collect-profiles-29319465-fptgb\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.294765 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f0797db-eac1-4318-8528-17c07458885a-config-volume\") pod \"collect-profiles-29319465-fptgb\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.294864 4667 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r6n9\" (UniqueName: \"kubernetes.io/projected/7f0797db-eac1-4318-8528-17c07458885a-kube-api-access-6r6n9\") pod \"collect-profiles-29319465-fptgb\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.396524 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r6n9\" (UniqueName: \"kubernetes.io/projected/7f0797db-eac1-4318-8528-17c07458885a-kube-api-access-6r6n9\") pod \"collect-profiles-29319465-fptgb\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.396609 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f0797db-eac1-4318-8528-17c07458885a-secret-volume\") pod \"collect-profiles-29319465-fptgb\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.396654 4667 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f0797db-eac1-4318-8528-17c07458885a-config-volume\") pod \"collect-profiles-29319465-fptgb\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.397387 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f0797db-eac1-4318-8528-17c07458885a-config-volume\") pod \"collect-profiles-29319465-fptgb\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.402013 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f0797db-eac1-4318-8528-17c07458885a-secret-volume\") pod \"collect-profiles-29319465-fptgb\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.409924 4667 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r6n9\" (UniqueName: \"kubernetes.io/projected/7f0797db-eac1-4318-8528-17c07458885a-kube-api-access-6r6n9\") pod \"collect-profiles-29319465-fptgb\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.445061 4667 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:00 crc kubenswrapper[4667]: I0929 17:45:00.801566 4667 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb"] Sep 29 17:45:01 crc kubenswrapper[4667]: I0929 17:45:01.710642 4667 generic.go:334] "Generic (PLEG): container finished" podID="7f0797db-eac1-4318-8528-17c07458885a" containerID="820ea1288e6f901f1c09539f5f2683b75180e3ea1e60280aee6ece8a9b153d68" exitCode=0 Sep 29 17:45:01 crc kubenswrapper[4667]: I0929 17:45:01.710746 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" event={"ID":"7f0797db-eac1-4318-8528-17c07458885a","Type":"ContainerDied","Data":"820ea1288e6f901f1c09539f5f2683b75180e3ea1e60280aee6ece8a9b153d68"} Sep 29 17:45:01 crc kubenswrapper[4667]: I0929 17:45:01.710859 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" event={"ID":"7f0797db-eac1-4318-8528-17c07458885a","Type":"ContainerStarted","Data":"9a93f36b22adffef5ac1b602759399398673e91a1e6baadf019fe2f3a8da748e"} Sep 29 17:45:02 crc kubenswrapper[4667]: I0929 17:45:02.943758 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.131267 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f0797db-eac1-4318-8528-17c07458885a-secret-volume\") pod \"7f0797db-eac1-4318-8528-17c07458885a\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.131412 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r6n9\" (UniqueName: \"kubernetes.io/projected/7f0797db-eac1-4318-8528-17c07458885a-kube-api-access-6r6n9\") pod \"7f0797db-eac1-4318-8528-17c07458885a\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.131526 4667 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f0797db-eac1-4318-8528-17c07458885a-config-volume\") pod \"7f0797db-eac1-4318-8528-17c07458885a\" (UID: \"7f0797db-eac1-4318-8528-17c07458885a\") " Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.132320 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f0797db-eac1-4318-8528-17c07458885a-config-volume" (OuterVolumeSpecName: "config-volume") pod "7f0797db-eac1-4318-8528-17c07458885a" (UID: "7f0797db-eac1-4318-8528-17c07458885a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.135292 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f0797db-eac1-4318-8528-17c07458885a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7f0797db-eac1-4318-8528-17c07458885a" (UID: "7f0797db-eac1-4318-8528-17c07458885a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.144128 4667 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f0797db-eac1-4318-8528-17c07458885a-kube-api-access-6r6n9" (OuterVolumeSpecName: "kube-api-access-6r6n9") pod "7f0797db-eac1-4318-8528-17c07458885a" (UID: "7f0797db-eac1-4318-8528-17c07458885a"). InnerVolumeSpecName "kube-api-access-6r6n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.233124 4667 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7f0797db-eac1-4318-8528-17c07458885a-secret-volume\") on node \"crc\" DevicePath \"\"" Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.233158 4667 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r6n9\" (UniqueName: \"kubernetes.io/projected/7f0797db-eac1-4318-8528-17c07458885a-kube-api-access-6r6n9\") on node \"crc\" DevicePath \"\"" Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.233171 4667 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f0797db-eac1-4318-8528-17c07458885a-config-volume\") on node \"crc\" DevicePath \"\"" Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.722151 4667 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" event={"ID":"7f0797db-eac1-4318-8528-17c07458885a","Type":"ContainerDied","Data":"9a93f36b22adffef5ac1b602759399398673e91a1e6baadf019fe2f3a8da748e"} Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.722185 4667 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a93f36b22adffef5ac1b602759399398673e91a1e6baadf019fe2f3a8da748e" Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.722201 4667 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29319465-fptgb" Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.992173 4667 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp"] Sep 29 17:45:03 crc kubenswrapper[4667]: I0929 17:45:03.996683 4667 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29319420-lnksp"] Sep 29 17:45:05 crc kubenswrapper[4667]: I0929 17:45:05.823447 4667 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68e21fe0-a72c-42e2-b2a8-89f8eed86bd2" path="/var/lib/kubelet/pods/68e21fe0-a72c-42e2-b2a8-89f8eed86bd2/volumes" Sep 29 17:45:08 crc kubenswrapper[4667]: E0929 17:45:08.817638 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:45:22 crc kubenswrapper[4667]: E0929 17:45:22.816910 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:45:26 crc kubenswrapper[4667]: I0929 17:45:26.897298 4667 patch_prober.go:28] interesting pod/machine-config-daemon-l8rmj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Sep 29 17:45:26 crc kubenswrapper[4667]: I0929 17:45:26.898183 4667 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-l8rmj" podUID="28fa0016-3e75-4704-8b60-30ee9e576d59" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Sep 29 17:45:32 crc kubenswrapper[4667]: I0929 17:45:32.411011 4667 scope.go:117] "RemoveContainer" containerID="43aea7ce1aff1605db00fce3ce3bb989b0242f18c2ff3dfd831faba070482a56" Sep 29 17:45:34 crc kubenswrapper[4667]: E0929 17:45:34.817138 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" Sep 29 17:45:45 crc kubenswrapper[4667]: E0929 17:45:45.817560 4667 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"38.129.56.34:5001/openstack-k8s-operators/openstack-operator-index:f45fa61af4e808ebd91514bf40d786904d9fd822\\\"\"" pod="openstack-operators/openstack-operator-index-6b7qn" podUID="7aa88991-5ef2-4166-8d37-8ce5fb207d26" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515066542723024457 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015066542723017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015066536301016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015066536301015462 5ustar corecore